Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
   1// SPDX-License-Identifier: GPL-2.0
   2
   3/*
   4 * Copyright 2016-2019 HabanaLabs, Ltd.
   5 * All Rights Reserved.
   6 */
   7
   8#include <uapi/misc/habanalabs.h>
   9#include "habanalabs.h"
  10#include "include/hw_ip/mmu/mmu_general.h"
  11
  12#include <linux/uaccess.h>
  13#include <linux/slab.h>
  14#include <linux/genalloc.h>
  15
  16#define PGS_IN_2MB_PAGE	(PAGE_SIZE_2MB >> PAGE_SHIFT)
  17#define HL_MMU_DEBUG	0
  18
  19/*
  20 * The va ranges in context object contain a list with the available chunks of
  21 * device virtual memory.
  22 * There is one range for host allocations and one for DRAM allocations.
  23 *
  24 * On initialization each range contains one chunk of all of its available
  25 * virtual range which is a half of the total device virtual range.
  26 *
  27 * On each mapping of physical pages, a suitable virtual range chunk (with a
  28 * minimum size) is selected from the list. If the chunk size equals the
  29 * requested size, the chunk is returned. Otherwise, the chunk is split into
  30 * two chunks - one to return as result and a remainder to stay in the list.
  31 *
  32 * On each Unmapping of a virtual address, the relevant virtual chunk is
  33 * returned to the list. The chunk is added to the list and if its edges match
  34 * the edges of the adjacent chunks (means a contiguous chunk can be created),
  35 * the chunks are merged.
  36 *
  37 * On finish, the list is checked to have only one chunk of all the relevant
  38 * virtual range (which is a half of the device total virtual range).
  39 * If not (means not all mappings were unmapped), a warning is printed.
  40 */
  41
  42/*
  43 * alloc_device_memory - allocate device memory
  44 *
  45 * @ctx                 : current context
  46 * @args                : host parameters containing the requested size
  47 * @ret_handle          : result handle
  48 *
  49 * This function does the following:
  50 * - Allocate the requested size rounded up to 2MB pages
  51 * - Return unique handle
  52 */
  53static int alloc_device_memory(struct hl_ctx *ctx, struct hl_mem_in *args,
  54				u32 *ret_handle)
  55{
  56	struct hl_device *hdev = ctx->hdev;
  57	struct hl_vm *vm = &hdev->vm;
  58	struct hl_vm_phys_pg_pack *phys_pg_pack;
  59	u64 paddr = 0, total_size, num_pgs, i;
  60	u32 num_curr_pgs, page_size, page_shift;
  61	int handle, rc;
  62	bool contiguous;
  63
  64	num_curr_pgs = 0;
  65	page_size = hdev->asic_prop.dram_page_size;
  66	page_shift = __ffs(page_size);
  67	num_pgs = (args->alloc.mem_size + (page_size - 1)) >> page_shift;
  68	total_size = num_pgs << page_shift;
  69
  70	contiguous = args->flags & HL_MEM_CONTIGUOUS;
  71
  72	if (contiguous) {
  73		paddr = (u64) gen_pool_alloc(vm->dram_pg_pool, total_size);
  74		if (!paddr) {
  75			dev_err(hdev->dev,
  76				"failed to allocate %llu huge contiguous pages\n",
  77				num_pgs);
  78			return -ENOMEM;
  79		}
  80	}
  81
  82	phys_pg_pack = kzalloc(sizeof(*phys_pg_pack), GFP_KERNEL);
  83	if (!phys_pg_pack) {
  84		rc = -ENOMEM;
  85		goto pages_pack_err;
  86	}
  87
  88	phys_pg_pack->vm_type = VM_TYPE_PHYS_PACK;
  89	phys_pg_pack->asid = ctx->asid;
  90	phys_pg_pack->npages = num_pgs;
  91	phys_pg_pack->page_size = page_size;
  92	phys_pg_pack->total_size = total_size;
  93	phys_pg_pack->flags = args->flags;
  94	phys_pg_pack->contiguous = contiguous;
  95
  96	phys_pg_pack->pages = kvmalloc_array(num_pgs, sizeof(u64), GFP_KERNEL);
  97	if (!phys_pg_pack->pages) {
  98		rc = -ENOMEM;
  99		goto pages_arr_err;
 100	}
 101
 102	if (phys_pg_pack->contiguous) {
 103		for (i = 0 ; i < num_pgs ; i++)
 104			phys_pg_pack->pages[i] = paddr + i * page_size;
 105	} else {
 106		for (i = 0 ; i < num_pgs ; i++) {
 107			phys_pg_pack->pages[i] = (u64) gen_pool_alloc(
 108							vm->dram_pg_pool,
 109							page_size);
 110			if (!phys_pg_pack->pages[i]) {
 111				dev_err(hdev->dev,
 112					"Failed to allocate device memory (out of memory)\n");
 113				rc = -ENOMEM;
 114				goto page_err;
 115			}
 116
 117			num_curr_pgs++;
 118		}
 119	}
 120
 121	spin_lock(&vm->idr_lock);
 122	handle = idr_alloc(&vm->phys_pg_pack_handles, phys_pg_pack, 1, 0,
 123				GFP_ATOMIC);
 124	spin_unlock(&vm->idr_lock);
 125
 126	if (handle < 0) {
 127		dev_err(hdev->dev, "Failed to get handle for page\n");
 128		rc = -EFAULT;
 129		goto idr_err;
 130	}
 131
 132	for (i = 0 ; i < num_pgs ; i++)
 133		kref_get(&vm->dram_pg_pool_refcount);
 134
 135	phys_pg_pack->handle = handle;
 136
 137	atomic64_add(phys_pg_pack->total_size, &ctx->dram_phys_mem);
 138	atomic64_add(phys_pg_pack->total_size, &hdev->dram_used_mem);
 139
 140	*ret_handle = handle;
 141
 142	return 0;
 143
 144idr_err:
 145page_err:
 146	if (!phys_pg_pack->contiguous)
 147		for (i = 0 ; i < num_curr_pgs ; i++)
 148			gen_pool_free(vm->dram_pg_pool, phys_pg_pack->pages[i],
 149					page_size);
 150
 151	kvfree(phys_pg_pack->pages);
 152pages_arr_err:
 153	kfree(phys_pg_pack);
 154pages_pack_err:
 155	if (contiguous)
 156		gen_pool_free(vm->dram_pg_pool, paddr, total_size);
 157
 158	return rc;
 159}
 160
 161/*
 162 * get_userptr_from_host_va - initialize userptr structure from given host
 163 *                            virtual address
 164 *
 165 * @hdev                : habanalabs device structure
 166 * @args                : parameters containing the virtual address and size
 167 * @p_userptr           : pointer to result userptr structure
 168 *
 169 * This function does the following:
 170 * - Allocate userptr structure
 171 * - Pin the given host memory using the userptr structure
 172 * - Perform DMA mapping to have the DMA addresses of the pages
 173 */
 174static int get_userptr_from_host_va(struct hl_device *hdev,
 175		struct hl_mem_in *args, struct hl_userptr **p_userptr)
 176{
 177	struct hl_userptr *userptr;
 178	int rc;
 179
 180	userptr = kzalloc(sizeof(*userptr), GFP_KERNEL);
 181	if (!userptr) {
 182		rc = -ENOMEM;
 183		goto userptr_err;
 184	}
 185
 186	rc = hl_pin_host_memory(hdev, args->map_host.host_virt_addr,
 187			args->map_host.mem_size, userptr);
 188	if (rc) {
 189		dev_err(hdev->dev, "Failed to pin host memory\n");
 190		goto pin_err;
 191	}
 192
 193	rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl,
 194					userptr->sgt->nents, DMA_BIDIRECTIONAL);
 195	if (rc) {
 196		dev_err(hdev->dev, "failed to map sgt with DMA region\n");
 197		goto dma_map_err;
 198	}
 199
 200	userptr->dma_mapped = true;
 201	userptr->dir = DMA_BIDIRECTIONAL;
 202	userptr->vm_type = VM_TYPE_USERPTR;
 203
 204	*p_userptr = userptr;
 205
 206	return 0;
 207
 208dma_map_err:
 209	hl_unpin_host_memory(hdev, userptr);
 210pin_err:
 211	kfree(userptr);
 212userptr_err:
 213
 214	return rc;
 215}
 216
 217/*
 218 * free_userptr - free userptr structure
 219 *
 220 * @hdev                : habanalabs device structure
 221 * @userptr             : userptr to free
 222 *
 223 * This function does the following:
 224 * - Unpins the physical pages
 225 * - Frees the userptr structure
 226 */
 227static void free_userptr(struct hl_device *hdev, struct hl_userptr *userptr)
 228{
 229	hl_unpin_host_memory(hdev, userptr);
 230	kfree(userptr);
 231}
 232
 233/*
 234 * dram_pg_pool_do_release - free DRAM pages pool
 235 *
 236 * @ref                 : pointer to reference object
 237 *
 238 * This function does the following:
 239 * - Frees the idr structure of physical pages handles
 240 * - Frees the generic pool of DRAM physical pages
 241 */
 242static void dram_pg_pool_do_release(struct kref *ref)
 243{
 244	struct hl_vm *vm = container_of(ref, struct hl_vm,
 245			dram_pg_pool_refcount);
 246
 247	/*
 248	 * free the idr here as only here we know for sure that there are no
 249	 * allocated physical pages and hence there are no handles in use
 250	 */
 251	idr_destroy(&vm->phys_pg_pack_handles);
 252	gen_pool_destroy(vm->dram_pg_pool);
 253}
 254
 255/*
 256 * free_phys_pg_pack   - free physical page pack
 257 *
 258 * @hdev               : habanalabs device structure
 259 * @phys_pg_pack       : physical page pack to free
 260 *
 261 * This function does the following:
 262 * - For DRAM memory only, iterate over the pack and free each physical block
 263 *   structure by returning it to the general pool
 264 * - Free the hl_vm_phys_pg_pack structure
 265 */
 266static void free_phys_pg_pack(struct hl_device *hdev,
 267		struct hl_vm_phys_pg_pack *phys_pg_pack)
 268{
 269	struct hl_vm *vm = &hdev->vm;
 270	u64 i;
 271
 272	if (!phys_pg_pack->created_from_userptr) {
 273		if (phys_pg_pack->contiguous) {
 274			gen_pool_free(vm->dram_pg_pool, phys_pg_pack->pages[0],
 275					phys_pg_pack->total_size);
 276
 277			for (i = 0; i < phys_pg_pack->npages ; i++)
 278				kref_put(&vm->dram_pg_pool_refcount,
 279					dram_pg_pool_do_release);
 280		} else {
 281			for (i = 0 ; i < phys_pg_pack->npages ; i++) {
 282				gen_pool_free(vm->dram_pg_pool,
 283						phys_pg_pack->pages[i],
 284						phys_pg_pack->page_size);
 285				kref_put(&vm->dram_pg_pool_refcount,
 286					dram_pg_pool_do_release);
 287			}
 288		}
 289	}
 290
 291	kvfree(phys_pg_pack->pages);
 292	kfree(phys_pg_pack);
 293}
 294
 295/*
 296 * free_device_memory - free device memory
 297 *
 298 * @ctx                  : current context
 299 * @handle              : handle of the memory chunk to free
 300 *
 301 * This function does the following:
 302 * - Free the device memory related to the given handle
 303 */
 304static int free_device_memory(struct hl_ctx *ctx, u32 handle)
 305{
 306	struct hl_device *hdev = ctx->hdev;
 307	struct hl_vm *vm = &hdev->vm;
 308	struct hl_vm_phys_pg_pack *phys_pg_pack;
 309
 310	spin_lock(&vm->idr_lock);
 311	phys_pg_pack = idr_find(&vm->phys_pg_pack_handles, handle);
 312	if (phys_pg_pack) {
 313		if (atomic_read(&phys_pg_pack->mapping_cnt) > 0) {
 314			dev_err(hdev->dev, "handle %u is mapped, cannot free\n",
 315				handle);
 316			spin_unlock(&vm->idr_lock);
 317			return -EINVAL;
 318		}
 319
 320		/*
 321		 * must remove from idr before the freeing of the physical
 322		 * pages as the refcount of the pool is also the trigger of the
 323		 * idr destroy
 324		 */
 325		idr_remove(&vm->phys_pg_pack_handles, handle);
 326		spin_unlock(&vm->idr_lock);
 327
 328		atomic64_sub(phys_pg_pack->total_size, &ctx->dram_phys_mem);
 329		atomic64_sub(phys_pg_pack->total_size, &hdev->dram_used_mem);
 330
 331		free_phys_pg_pack(hdev, phys_pg_pack);
 332	} else {
 333		spin_unlock(&vm->idr_lock);
 334		dev_err(hdev->dev,
 335			"free device memory failed, no match for handle %u\n",
 336			handle);
 337		return -EINVAL;
 338	}
 339
 340	return 0;
 341}
 342
 343/*
 344 * clear_va_list_locked - free virtual addresses list
 345 *
 346 * @hdev                : habanalabs device structure
 347 * @va_list             : list of virtual addresses to free
 348 *
 349 * This function does the following:
 350 * - Iterate over the list and free each virtual addresses block
 351 *
 352 * This function should be called only when va_list lock is taken
 353 */
 354static void clear_va_list_locked(struct hl_device *hdev,
 355		struct list_head *va_list)
 356{
 357	struct hl_vm_va_block *va_block, *tmp;
 358
 359	list_for_each_entry_safe(va_block, tmp, va_list, node) {
 360		list_del(&va_block->node);
 361		kfree(va_block);
 362	}
 363}
 364
 365/*
 366 * print_va_list_locked    - print virtual addresses list
 367 *
 368 * @hdev                : habanalabs device structure
 369 * @va_list             : list of virtual addresses to print
 370 *
 371 * This function does the following:
 372 * - Iterate over the list and print each virtual addresses block
 373 *
 374 * This function should be called only when va_list lock is taken
 375 */
 376static void print_va_list_locked(struct hl_device *hdev,
 377		struct list_head *va_list)
 378{
 379#if HL_MMU_DEBUG
 380	struct hl_vm_va_block *va_block;
 381
 382	dev_dbg(hdev->dev, "print va list:\n");
 383
 384	list_for_each_entry(va_block, va_list, node)
 385		dev_dbg(hdev->dev,
 386			"va block, start: 0x%llx, end: 0x%llx, size: %llu\n",
 387			va_block->start, va_block->end, va_block->size);
 388#endif
 389}
 390
 391/*
 392 * merge_va_blocks_locked - merge a virtual block if possible
 393 *
 394 * @hdev                : pointer to the habanalabs device structure
 395 * @va_list             : pointer to the virtual addresses block list
 396 * @va_block            : virtual block to merge with adjacent blocks
 397 *
 398 * This function does the following:
 399 * - Merge the given blocks with the adjacent blocks if their virtual ranges
 400 *   create a contiguous virtual range
 401 *
 402 * This Function should be called only when va_list lock is taken
 403 */
 404static void merge_va_blocks_locked(struct hl_device *hdev,
 405		struct list_head *va_list, struct hl_vm_va_block *va_block)
 406{
 407	struct hl_vm_va_block *prev, *next;
 408
 409	prev = list_prev_entry(va_block, node);
 410	if (&prev->node != va_list && prev->end + 1 == va_block->start) {
 411		prev->end = va_block->end;
 412		prev->size = prev->end - prev->start;
 413		list_del(&va_block->node);
 414		kfree(va_block);
 415		va_block = prev;
 416	}
 417
 418	next = list_next_entry(va_block, node);
 419	if (&next->node != va_list && va_block->end + 1 == next->start) {
 420		next->start = va_block->start;
 421		next->size = next->end - next->start;
 422		list_del(&va_block->node);
 423		kfree(va_block);
 424	}
 425}
 426
 427/*
 428 * add_va_block_locked - add a virtual block to the virtual addresses list
 429 *
 430 * @hdev                : pointer to the habanalabs device structure
 431 * @va_list             : pointer to the virtual addresses block list
 432 * @start               : start virtual address
 433 * @end                 : end virtual address
 434 *
 435 * This function does the following:
 436 * - Add the given block to the virtual blocks list and merge with other
 437 * blocks if a contiguous virtual block can be created
 438 *
 439 * This Function should be called only when va_list lock is taken
 440 */
 441static int add_va_block_locked(struct hl_device *hdev,
 442		struct list_head *va_list, u64 start, u64 end)
 443{
 444	struct hl_vm_va_block *va_block, *res = NULL;
 445	u64 size = end - start;
 446
 447	print_va_list_locked(hdev, va_list);
 448
 449	list_for_each_entry(va_block, va_list, node) {
 450		/* TODO: remove upon matureness */
 451		if (hl_mem_area_crosses_range(start, size, va_block->start,
 452				va_block->end)) {
 453			dev_err(hdev->dev,
 454				"block crossing ranges at start 0x%llx, end 0x%llx\n",
 455				va_block->start, va_block->end);
 456			return -EINVAL;
 457		}
 458
 459		if (va_block->end < start)
 460			res = va_block;
 461	}
 462
 463	va_block = kmalloc(sizeof(*va_block), GFP_KERNEL);
 464	if (!va_block)
 465		return -ENOMEM;
 466
 467	va_block->start = start;
 468	va_block->end = end;
 469	va_block->size = size;
 470
 471	if (!res)
 472		list_add(&va_block->node, va_list);
 473	else
 474		list_add(&va_block->node, &res->node);
 475
 476	merge_va_blocks_locked(hdev, va_list, va_block);
 477
 478	print_va_list_locked(hdev, va_list);
 479
 480	return 0;
 481}
 482
 483/*
 484 * add_va_block - wrapper for add_va_block_locked
 485 *
 486 * @hdev                : pointer to the habanalabs device structure
 487 * @va_list             : pointer to the virtual addresses block list
 488 * @start               : start virtual address
 489 * @end                 : end virtual address
 490 *
 491 * This function does the following:
 492 * - Takes the list lock and calls add_va_block_locked
 493 */
 494static inline int add_va_block(struct hl_device *hdev,
 495		struct hl_va_range *va_range, u64 start, u64 end)
 496{
 497	int rc;
 498
 499	mutex_lock(&va_range->lock);
 500	rc = add_va_block_locked(hdev, &va_range->list, start, end);
 501	mutex_unlock(&va_range->lock);
 502
 503	return rc;
 504}
 505
 506/*
 507 * get_va_block - get a virtual block with the requested size
 508 *
 509 * @hdev            : pointer to the habanalabs device structure
 510 * @va_range        : pointer to the virtual addresses range
 511 * @size            : requested block size
 512 * @hint_addr       : hint for request address by the user
 513 * @is_userptr      : is host or DRAM memory
 514 *
 515 * This function does the following:
 516 * - Iterate on the virtual block list to find a suitable virtual block for the
 517 *   requested size
 518 * - Reserve the requested block and update the list
 519 * - Return the start address of the virtual block
 520 */
 521static u64 get_va_block(struct hl_device *hdev,
 522		struct hl_va_range *va_range, u64 size, u64 hint_addr,
 523		bool is_userptr)
 524{
 525	struct hl_vm_va_block *va_block, *new_va_block = NULL;
 526	u64 valid_start, valid_size, prev_start, prev_end, page_mask,
 527		res_valid_start = 0, res_valid_size = 0;
 528	u32 page_size;
 529	bool add_prev = false;
 530
 531	if (is_userptr) {
 532		/*
 533		 * We cannot know if the user allocated memory with huge pages
 534		 * or not, hence we continue with the biggest possible
 535		 * granularity.
 536		 */
 537		page_size = PAGE_SIZE_2MB;
 538		page_mask = PAGE_MASK_2MB;
 539	} else {
 540		page_size = hdev->asic_prop.dram_page_size;
 541		page_mask = ~((u64)page_size - 1);
 542	}
 543
 544	mutex_lock(&va_range->lock);
 545
 546	print_va_list_locked(hdev, &va_range->list);
 547
 548	list_for_each_entry(va_block, &va_range->list, node) {
 549		/* calc the first possible aligned addr */
 550		valid_start = va_block->start;
 551
 552
 553		if (valid_start & (page_size - 1)) {
 554			valid_start &= page_mask;
 555			valid_start += page_size;
 556			if (valid_start > va_block->end)
 557				continue;
 558		}
 559
 560		valid_size = va_block->end - valid_start;
 561
 562		if (valid_size >= size &&
 563			(!new_va_block || valid_size < res_valid_size)) {
 564
 565			new_va_block = va_block;
 566			res_valid_start = valid_start;
 567			res_valid_size = valid_size;
 568		}
 569
 570		if (hint_addr && hint_addr >= valid_start &&
 571				((hint_addr + size) <= va_block->end)) {
 572			new_va_block = va_block;
 573			res_valid_start = hint_addr;
 574			res_valid_size = valid_size;
 575			break;
 576		}
 577	}
 578
 579	if (!new_va_block) {
 580		dev_err(hdev->dev, "no available va block for size %llu\n",
 581				size);
 582		goto out;
 583	}
 584
 585	if (res_valid_start > new_va_block->start) {
 586		prev_start = new_va_block->start;
 587		prev_end = res_valid_start - 1;
 588
 589		new_va_block->start = res_valid_start;
 590		new_va_block->size = res_valid_size;
 591
 592		add_prev = true;
 593	}
 594
 595	if (new_va_block->size > size) {
 596		new_va_block->start += size;
 597		new_va_block->size = new_va_block->end - new_va_block->start;
 598	} else {
 599		list_del(&new_va_block->node);
 600		kfree(new_va_block);
 601	}
 602
 603	if (add_prev)
 604		add_va_block_locked(hdev, &va_range->list, prev_start,
 605				prev_end);
 606
 607	print_va_list_locked(hdev, &va_range->list);
 608out:
 609	mutex_unlock(&va_range->lock);
 610
 611	return res_valid_start;
 612}
 613
 614/*
 615 * get_sg_info - get number of pages and the DMA address from SG list
 616 *
 617 * @sg                 : the SG list
 618 * @dma_addr           : pointer to DMA address to return
 619 *
 620 * Calculate the number of consecutive pages described by the SG list. Take the
 621 * offset of the address in the first page, add to it the length and round it up
 622 * to the number of needed pages.
 623 */
 624static u32 get_sg_info(struct scatterlist *sg, dma_addr_t *dma_addr)
 625{
 626	*dma_addr = sg_dma_address(sg);
 627
 628	return ((((*dma_addr) & (PAGE_SIZE - 1)) + sg_dma_len(sg)) +
 629			(PAGE_SIZE - 1)) >> PAGE_SHIFT;
 630}
 631
 632/*
 633 * init_phys_pg_pack_from_userptr - initialize physical page pack from host
 634 *                                   memory
 635 *
 636 * @ctx                : current context
 637 * @userptr            : userptr to initialize from
 638 * @pphys_pg_pack      : res pointer
 639 *
 640 * This function does the following:
 641 * - Pin the physical pages related to the given virtual block
 642 * - Create a physical page pack from the physical pages related to the given
 643 *   virtual block
 644 */
 645static int init_phys_pg_pack_from_userptr(struct hl_ctx *ctx,
 646		struct hl_userptr *userptr,
 647		struct hl_vm_phys_pg_pack **pphys_pg_pack)
 648{
 649	struct hl_vm_phys_pg_pack *phys_pg_pack;
 650	struct scatterlist *sg;
 651	dma_addr_t dma_addr;
 652	u64 page_mask, total_npages;
 653	u32 npages, page_size = PAGE_SIZE;
 654	bool first = true, is_huge_page_opt = true;
 655	int rc, i, j;
 656
 657	phys_pg_pack = kzalloc(sizeof(*phys_pg_pack), GFP_KERNEL);
 658	if (!phys_pg_pack)
 659		return -ENOMEM;
 660
 661	phys_pg_pack->vm_type = userptr->vm_type;
 662	phys_pg_pack->created_from_userptr = true;
 663	phys_pg_pack->asid = ctx->asid;
 664	atomic_set(&phys_pg_pack->mapping_cnt, 1);
 665
 666	/* Only if all dma_addrs are aligned to 2MB and their
 667	 * sizes is at least 2MB, we can use huge page mapping.
 668	 * We limit the 2MB optimization to this condition,
 669	 * since later on we acquire the related VA range as one
 670	 * consecutive block.
 671	 */
 672	total_npages = 0;
 673	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
 674		npages = get_sg_info(sg, &dma_addr);
 675
 676		total_npages += npages;
 677
 678		if ((npages % PGS_IN_2MB_PAGE) ||
 679					(dma_addr & (PAGE_SIZE_2MB - 1)))
 680			is_huge_page_opt = false;
 681	}
 682
 683	if (is_huge_page_opt) {
 684		page_size = PAGE_SIZE_2MB;
 685		total_npages /= PGS_IN_2MB_PAGE;
 686	}
 687
 688	page_mask = ~(((u64) page_size) - 1);
 689
 690	phys_pg_pack->pages = kvmalloc_array(total_npages, sizeof(u64),
 691						GFP_KERNEL);
 692	if (!phys_pg_pack->pages) {
 693		rc = -ENOMEM;
 694		goto page_pack_arr_mem_err;
 695	}
 696
 697	phys_pg_pack->npages = total_npages;
 698	phys_pg_pack->page_size = page_size;
 699	phys_pg_pack->total_size = total_npages * page_size;
 700
 701	j = 0;
 702	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
 703		npages = get_sg_info(sg, &dma_addr);
 704
 705		/* align down to physical page size and save the offset */
 706		if (first) {
 707			first = false;
 708			phys_pg_pack->offset = dma_addr & (page_size - 1);
 709			dma_addr &= page_mask;
 710		}
 711
 712		while (npages) {
 713			phys_pg_pack->pages[j++] = dma_addr;
 714			dma_addr += page_size;
 715
 716			if (is_huge_page_opt)
 717				npages -= PGS_IN_2MB_PAGE;
 718			else
 719				npages--;
 720		}
 721	}
 722
 723	*pphys_pg_pack = phys_pg_pack;
 724
 725	return 0;
 726
 727page_pack_arr_mem_err:
 728	kfree(phys_pg_pack);
 729
 730	return rc;
 731}
 732
 733/*
 734 * map_phys_page_pack - maps the physical page pack
 735 *
 736 * @ctx                : current context
 737 * @vaddr              : start address of the virtual area to map from
 738 * @phys_pg_pack       : the pack of physical pages to map to
 739 *
 740 * This function does the following:
 741 * - Maps each chunk of virtual memory to matching physical chunk
 742 * - Stores number of successful mappings in the given argument
 743 * - Returns 0 on success, error code otherwise.
 744 */
 745static int map_phys_page_pack(struct hl_ctx *ctx, u64 vaddr,
 746		struct hl_vm_phys_pg_pack *phys_pg_pack)
 747{
 748	struct hl_device *hdev = ctx->hdev;
 749	u64 next_vaddr = vaddr, paddr, mapped_pg_cnt = 0, i;
 750	u32 page_size = phys_pg_pack->page_size;
 751	int rc = 0;
 752
 753	for (i = 0 ; i < phys_pg_pack->npages ; i++) {
 754		paddr = phys_pg_pack->pages[i];
 755
 756		rc = hl_mmu_map(ctx, next_vaddr, paddr, page_size);
 757		if (rc) {
 758			dev_err(hdev->dev,
 759				"map failed for handle %u, npages: %llu, mapped: %llu",
 760				phys_pg_pack->handle, phys_pg_pack->npages,
 761				mapped_pg_cnt);
 762			goto err;
 763		}
 764
 765		mapped_pg_cnt++;
 766		next_vaddr += page_size;
 767	}
 768
 769	return 0;
 770
 771err:
 772	next_vaddr = vaddr;
 773	for (i = 0 ; i < mapped_pg_cnt ; i++) {
 774		if (hl_mmu_unmap(ctx, next_vaddr, page_size))
 775			dev_warn_ratelimited(hdev->dev,
 776				"failed to unmap handle %u, va: 0x%llx, pa: 0x%llx, page size: %u\n",
 777					phys_pg_pack->handle, next_vaddr,
 778					phys_pg_pack->pages[i], page_size);
 779
 780		next_vaddr += page_size;
 781	}
 782
 783	return rc;
 784}
 785
 786static int get_paddr_from_handle(struct hl_ctx *ctx, struct hl_mem_in *args,
 787				u64 *paddr)
 788{
 789	struct hl_device *hdev = ctx->hdev;
 790	struct hl_vm *vm = &hdev->vm;
 791	struct hl_vm_phys_pg_pack *phys_pg_pack;
 792	u32 handle;
 793
 794	handle = lower_32_bits(args->map_device.handle);
 795	spin_lock(&vm->idr_lock);
 796	phys_pg_pack = idr_find(&vm->phys_pg_pack_handles, handle);
 797	if (!phys_pg_pack) {
 798		spin_unlock(&vm->idr_lock);
 799		dev_err(hdev->dev, "no match for handle %u\n", handle);
 800		return -EINVAL;
 801	}
 802
 803	*paddr = phys_pg_pack->pages[0];
 804
 805	spin_unlock(&vm->idr_lock);
 806
 807	return 0;
 808}
 809
 810/*
 811 * map_device_va - map the given memory
 812 *
 813 * @ctx	         : current context
 814 * @args         : host parameters with handle/host virtual address
 815 * @device_addr	 : pointer to result device virtual address
 816 *
 817 * This function does the following:
 818 * - If given a physical device memory handle, map to a device virtual block
 819 *   and return the start address of this block
 820 * - If given a host virtual address and size, find the related physical pages,
 821 *   map a device virtual block to this pages and return the start address of
 822 *   this block
 823 */
 824static int map_device_va(struct hl_ctx *ctx, struct hl_mem_in *args,
 825		u64 *device_addr)
 826{
 827	struct hl_device *hdev = ctx->hdev;
 828	struct hl_vm *vm = &hdev->vm;
 829	struct hl_vm_phys_pg_pack *phys_pg_pack;
 830	struct hl_userptr *userptr = NULL;
 831	struct hl_vm_hash_node *hnode;
 832	enum vm_type_t *vm_type;
 833	u64 ret_vaddr, hint_addr;
 834	u32 handle = 0;
 835	int rc;
 836	bool is_userptr = args->flags & HL_MEM_USERPTR;
 837
 838	/* Assume failure */
 839	*device_addr = 0;
 840
 841	if (is_userptr) {
 842		rc = get_userptr_from_host_va(hdev, args, &userptr);
 843		if (rc) {
 844			dev_err(hdev->dev, "failed to get userptr from va\n");
 845			return rc;
 846		}
 847
 848		rc = init_phys_pg_pack_from_userptr(ctx, userptr,
 849				&phys_pg_pack);
 850		if (rc) {
 851			dev_err(hdev->dev,
 852				"unable to init page pack for vaddr 0x%llx\n",
 853				args->map_host.host_virt_addr);
 854			goto init_page_pack_err;
 855		}
 856
 857		vm_type = (enum vm_type_t *) userptr;
 858		hint_addr = args->map_host.hint_addr;
 859	} else {
 860		handle = lower_32_bits(args->map_device.handle);
 861
 862		spin_lock(&vm->idr_lock);
 863		phys_pg_pack = idr_find(&vm->phys_pg_pack_handles, handle);
 864		if (!phys_pg_pack) {
 865			spin_unlock(&vm->idr_lock);
 866			dev_err(hdev->dev,
 867				"no match for handle %u\n", handle);
 868			return -EINVAL;
 869		}
 870
 871		/* increment now to avoid freeing device memory while mapping */
 872		atomic_inc(&phys_pg_pack->mapping_cnt);
 873
 874		spin_unlock(&vm->idr_lock);
 875
 876		vm_type = (enum vm_type_t *) phys_pg_pack;
 877
 878		hint_addr = args->map_device.hint_addr;
 879	}
 880
 881	/*
 882	 * relevant for mapping device physical memory only, as host memory is
 883	 * implicitly shared
 884	 */
 885	if (!is_userptr && !(phys_pg_pack->flags & HL_MEM_SHARED) &&
 886			phys_pg_pack->asid != ctx->asid) {
 887		dev_err(hdev->dev,
 888			"Failed to map memory, handle %u is not shared\n",
 889			handle);
 890		rc = -EPERM;
 891		goto shared_err;
 892	}
 893
 894	hnode = kzalloc(sizeof(*hnode), GFP_KERNEL);
 895	if (!hnode) {
 896		rc = -ENOMEM;
 897		goto hnode_err;
 898	}
 899
 900	ret_vaddr = get_va_block(hdev,
 901			is_userptr ? &ctx->host_va_range : &ctx->dram_va_range,
 902			phys_pg_pack->total_size, hint_addr, is_userptr);
 903	if (!ret_vaddr) {
 904		dev_err(hdev->dev, "no available va block for handle %u\n",
 905				handle);
 906		rc = -ENOMEM;
 907		goto va_block_err;
 908	}
 909
 910	mutex_lock(&ctx->mmu_lock);
 911
 912	rc = map_phys_page_pack(ctx, ret_vaddr, phys_pg_pack);
 913	if (rc) {
 914		mutex_unlock(&ctx->mmu_lock);
 915		dev_err(hdev->dev, "mapping page pack failed for handle %u\n",
 916				handle);
 917		goto map_err;
 918	}
 919
 920	hdev->asic_funcs->mmu_invalidate_cache(hdev, false);
 921
 922	mutex_unlock(&ctx->mmu_lock);
 923
 924	ret_vaddr += phys_pg_pack->offset;
 925
 926	hnode->ptr = vm_type;
 927	hnode->vaddr = ret_vaddr;
 928
 929	mutex_lock(&ctx->mem_hash_lock);
 930	hash_add(ctx->mem_hash, &hnode->node, ret_vaddr);
 931	mutex_unlock(&ctx->mem_hash_lock);
 932
 933	*device_addr = ret_vaddr;
 934
 935	if (is_userptr)
 936		free_phys_pg_pack(hdev, phys_pg_pack);
 937
 938	return 0;
 939
 940map_err:
 941	if (add_va_block(hdev,
 942			is_userptr ? &ctx->host_va_range : &ctx->dram_va_range,
 943			ret_vaddr,
 944			ret_vaddr + phys_pg_pack->total_size - 1))
 945		dev_warn(hdev->dev,
 946			"release va block failed for handle 0x%x, vaddr: 0x%llx\n",
 947				handle, ret_vaddr);
 948
 949va_block_err:
 950	kfree(hnode);
 951hnode_err:
 952shared_err:
 953	atomic_dec(&phys_pg_pack->mapping_cnt);
 954	if (is_userptr)
 955		free_phys_pg_pack(hdev, phys_pg_pack);
 956init_page_pack_err:
 957	if (is_userptr)
 958		free_userptr(hdev, userptr);
 959
 960	return rc;
 961}
 962
 963/*
 964 * unmap_device_va      - unmap the given device virtual address
 965 *
 966 * @ctx                 : current context
 967 * @vaddr               : device virtual address to unmap
 968 *
 969 * This function does the following:
 970 * - Unmap the physical pages related to the given virtual address
 971 * - return the device virtual block to the virtual block list
 972 */
 973static int unmap_device_va(struct hl_ctx *ctx, u64 vaddr)
 974{
 975	struct hl_device *hdev = ctx->hdev;
 976	struct hl_vm_phys_pg_pack *phys_pg_pack = NULL;
 977	struct hl_vm_hash_node *hnode = NULL;
 978	struct hl_userptr *userptr = NULL;
 979	enum vm_type_t *vm_type;
 980	u64 next_vaddr, i;
 981	u32 page_size;
 982	bool is_userptr;
 983	int rc;
 984
 985	/* protect from double entrance */
 986	mutex_lock(&ctx->mem_hash_lock);
 987	hash_for_each_possible(ctx->mem_hash, hnode, node, (unsigned long)vaddr)
 988		if (vaddr == hnode->vaddr)
 989			break;
 990
 991	if (!hnode) {
 992		mutex_unlock(&ctx->mem_hash_lock);
 993		dev_err(hdev->dev,
 994			"unmap failed, no mem hnode for vaddr 0x%llx\n",
 995			vaddr);
 996		return -EINVAL;
 997	}
 998
 999	hash_del(&hnode->node);
1000	mutex_unlock(&ctx->mem_hash_lock);
1001
1002	vm_type = hnode->ptr;
1003
1004	if (*vm_type == VM_TYPE_USERPTR) {
1005		is_userptr = true;
1006		userptr = hnode->ptr;
1007		rc = init_phys_pg_pack_from_userptr(ctx, userptr,
1008				&phys_pg_pack);
1009		if (rc) {
1010			dev_err(hdev->dev,
1011				"unable to init page pack for vaddr 0x%llx\n",
1012				vaddr);
1013			goto vm_type_err;
1014		}
1015	} else if (*vm_type == VM_TYPE_PHYS_PACK) {
1016		is_userptr = false;
1017		phys_pg_pack = hnode->ptr;
1018	} else {
1019		dev_warn(hdev->dev,
1020			"unmap failed, unknown vm desc for vaddr 0x%llx\n",
1021				vaddr);
1022		rc = -EFAULT;
1023		goto vm_type_err;
1024	}
1025
1026	if (atomic_read(&phys_pg_pack->mapping_cnt) == 0) {
1027		dev_err(hdev->dev, "vaddr 0x%llx is not mapped\n", vaddr);
1028		rc = -EINVAL;
1029		goto mapping_cnt_err;
1030	}
1031
1032	page_size = phys_pg_pack->page_size;
1033	vaddr &= ~(((u64) page_size) - 1);
1034
1035	next_vaddr = vaddr;
1036
1037	mutex_lock(&ctx->mmu_lock);
1038
1039	for (i = 0 ; i < phys_pg_pack->npages ; i++, next_vaddr += page_size) {
1040		if (hl_mmu_unmap(ctx, next_vaddr, page_size))
1041			dev_warn_ratelimited(hdev->dev,
1042			"unmap failed for vaddr: 0x%llx\n", next_vaddr);
1043
1044		/* unmapping on Palladium can be really long, so avoid a CPU
1045		 * soft lockup bug by sleeping a little between unmapping pages
1046		 */
1047		if (hdev->pldm)
1048			usleep_range(500, 1000);
1049	}
1050
1051	hdev->asic_funcs->mmu_invalidate_cache(hdev, true);
1052
1053	mutex_unlock(&ctx->mmu_lock);
1054
1055	if (add_va_block(hdev,
1056			is_userptr ? &ctx->host_va_range : &ctx->dram_va_range,
1057			vaddr,
1058			vaddr + phys_pg_pack->total_size - 1))
1059		dev_warn(hdev->dev, "add va block failed for vaddr: 0x%llx\n",
1060				vaddr);
1061
1062	atomic_dec(&phys_pg_pack->mapping_cnt);
1063	kfree(hnode);
1064
1065	if (is_userptr) {
1066		free_phys_pg_pack(hdev, phys_pg_pack);
1067		free_userptr(hdev, userptr);
1068	}
1069
1070	return 0;
1071
1072mapping_cnt_err:
1073	if (is_userptr)
1074		free_phys_pg_pack(hdev, phys_pg_pack);
1075vm_type_err:
1076	mutex_lock(&ctx->mem_hash_lock);
1077	hash_add(ctx->mem_hash, &hnode->node, vaddr);
1078	mutex_unlock(&ctx->mem_hash_lock);
1079
1080	return rc;
1081}
1082
1083static int mem_ioctl_no_mmu(struct hl_fpriv *hpriv, union hl_mem_args *args)
1084{
1085	struct hl_device *hdev = hpriv->hdev;
1086	struct hl_ctx *ctx = hpriv->ctx;
1087	u64 device_addr = 0;
1088	u32 handle = 0;
1089	int rc;
1090
1091	switch (args->in.op) {
1092	case HL_MEM_OP_ALLOC:
1093		if (args->in.alloc.mem_size == 0) {
1094			dev_err(hdev->dev,
1095				"alloc size must be larger than 0\n");
1096			rc = -EINVAL;
1097			goto out;
1098		}
1099
1100		/* Force contiguous as there are no real MMU
1101		 * translations to overcome physical memory gaps
1102		 */
1103		args->in.flags |= HL_MEM_CONTIGUOUS;
1104		rc = alloc_device_memory(ctx, &args->in, &handle);
1105
1106		memset(args, 0, sizeof(*args));
1107		args->out.handle = (__u64) handle;
1108		break;
1109
1110	case HL_MEM_OP_FREE:
1111		rc = free_device_memory(ctx, args->in.free.handle);
1112		break;
1113
1114	case HL_MEM_OP_MAP:
1115		if (args->in.flags & HL_MEM_USERPTR) {
1116			device_addr = args->in.map_host.host_virt_addr;
1117			rc = 0;
1118		} else {
1119			rc = get_paddr_from_handle(ctx, &args->in,
1120					&device_addr);
1121		}
1122
1123		memset(args, 0, sizeof(*args));
1124		args->out.device_virt_addr = device_addr;
1125		break;
1126
1127	case HL_MEM_OP_UNMAP:
1128		rc = 0;
1129		break;
1130
1131	default:
1132		dev_err(hdev->dev, "Unknown opcode for memory IOCTL\n");
1133		rc = -ENOTTY;
1134		break;
1135	}
1136
1137out:
1138	return rc;
1139}
1140
1141int hl_mem_ioctl(struct hl_fpriv *hpriv, void *data)
1142{
1143	union hl_mem_args *args = data;
1144	struct hl_device *hdev = hpriv->hdev;
1145	struct hl_ctx *ctx = hpriv->ctx;
1146	u64 device_addr = 0;
1147	u32 handle = 0;
1148	int rc;
1149
1150	if (hl_device_disabled_or_in_reset(hdev)) {
1151		dev_warn_ratelimited(hdev->dev,
1152			"Device is %s. Can't execute MEMORY IOCTL\n",
1153			atomic_read(&hdev->in_reset) ? "in_reset" : "disabled");
1154		return -EBUSY;
1155	}
1156
1157	if (!hdev->mmu_enable)
1158		return mem_ioctl_no_mmu(hpriv, args);
1159
1160	switch (args->in.op) {
1161	case HL_MEM_OP_ALLOC:
1162		if (!hdev->dram_supports_virtual_memory) {
1163			dev_err(hdev->dev, "DRAM alloc is not supported\n");
1164			rc = -EINVAL;
1165			goto out;
1166		}
1167
1168		if (args->in.alloc.mem_size == 0) {
1169			dev_err(hdev->dev,
1170				"alloc size must be larger than 0\n");
1171			rc = -EINVAL;
1172			goto out;
1173		}
1174		rc = alloc_device_memory(ctx, &args->in, &handle);
1175
1176		memset(args, 0, sizeof(*args));
1177		args->out.handle = (__u64) handle;
1178		break;
1179
1180	case HL_MEM_OP_FREE:
1181		rc = free_device_memory(ctx, args->in.free.handle);
1182		break;
1183
1184	case HL_MEM_OP_MAP:
1185		rc = map_device_va(ctx, &args->in, &device_addr);
1186
1187		memset(args, 0, sizeof(*args));
1188		args->out.device_virt_addr = device_addr;
1189		break;
1190
1191	case HL_MEM_OP_UNMAP:
1192		rc = unmap_device_va(ctx,
1193				args->in.unmap.device_virt_addr);
1194		break;
1195
1196	default:
1197		dev_err(hdev->dev, "Unknown opcode for memory IOCTL\n");
1198		rc = -ENOTTY;
1199		break;
1200	}
1201
1202out:
1203	return rc;
1204}
1205
1206/*
1207 * hl_pin_host_memory - pins a chunk of host memory
1208 *
1209 * @hdev                : pointer to the habanalabs device structure
1210 * @addr                : the user-space virtual address of the memory area
1211 * @size                : the size of the memory area
1212 * @userptr	        : pointer to hl_userptr structure
1213 *
1214 * This function does the following:
1215 * - Pins the physical pages
1216 * - Create a SG list from those pages
1217 */
1218int hl_pin_host_memory(struct hl_device *hdev, u64 addr, u64 size,
1219			struct hl_userptr *userptr)
1220{
1221	u64 start, end;
1222	u32 npages, offset;
1223	int rc;
1224
1225	if (!size) {
1226		dev_err(hdev->dev, "size to pin is invalid - %llu\n", size);
1227		return -EINVAL;
1228	}
1229
1230	if (!access_ok((void __user *) (uintptr_t) addr, size)) {
1231		dev_err(hdev->dev, "user pointer is invalid - 0x%llx\n", addr);
1232		return -EFAULT;
1233	}
1234
1235	/*
1236	 * If the combination of the address and size requested for this memory
1237	 * region causes an integer overflow, return error.
1238	 */
1239	if (((addr + size) < addr) ||
1240			PAGE_ALIGN(addr + size) < (addr + size)) {
1241		dev_err(hdev->dev,
1242			"user pointer 0x%llx + %llu causes integer overflow\n",
1243			addr, size);
1244		return -EINVAL;
1245	}
1246
1247	start = addr & PAGE_MASK;
1248	offset = addr & ~PAGE_MASK;
1249	end = PAGE_ALIGN(addr + size);
1250	npages = (end - start) >> PAGE_SHIFT;
1251
1252	userptr->size = size;
1253	userptr->addr = addr;
1254	userptr->dma_mapped = false;
1255	INIT_LIST_HEAD(&userptr->job_node);
1256
1257	userptr->vec = frame_vector_create(npages);
1258	if (!userptr->vec) {
1259		dev_err(hdev->dev, "Failed to create frame vector\n");
1260		return -ENOMEM;
1261	}
1262
1263	rc = get_vaddr_frames(start, npages, FOLL_FORCE | FOLL_WRITE,
1264				userptr->vec);
1265
1266	if (rc != npages) {
1267		dev_err(hdev->dev,
1268			"Failed to map host memory, user ptr probably wrong\n");
1269		if (rc < 0)
1270			goto destroy_framevec;
1271		rc = -EFAULT;
1272		goto put_framevec;
1273	}
1274
1275	if (frame_vector_to_pages(userptr->vec) < 0) {
1276		dev_err(hdev->dev,
1277			"Failed to translate frame vector to pages\n");
1278		rc = -EFAULT;
1279		goto put_framevec;
1280	}
1281
1282	userptr->sgt = kzalloc(sizeof(*userptr->sgt), GFP_ATOMIC);
1283	if (!userptr->sgt) {
1284		rc = -ENOMEM;
1285		goto put_framevec;
1286	}
1287
1288	rc = sg_alloc_table_from_pages(userptr->sgt,
1289					frame_vector_pages(userptr->vec),
1290					npages, offset, size, GFP_ATOMIC);
1291	if (rc < 0) {
1292		dev_err(hdev->dev, "failed to create SG table from pages\n");
1293		goto free_sgt;
1294	}
1295
1296	hl_debugfs_add_userptr(hdev, userptr);
1297
1298	return 0;
1299
1300free_sgt:
1301	kfree(userptr->sgt);
1302put_framevec:
1303	put_vaddr_frames(userptr->vec);
1304destroy_framevec:
1305	frame_vector_destroy(userptr->vec);
1306	return rc;
1307}
1308
1309/*
1310 * hl_unpin_host_memory - unpins a chunk of host memory
1311 *
1312 * @hdev                : pointer to the habanalabs device structure
1313 * @userptr             : pointer to hl_userptr structure
1314 *
1315 * This function does the following:
1316 * - Unpins the physical pages related to the host memory
1317 * - Free the SG list
1318 */
1319int hl_unpin_host_memory(struct hl_device *hdev, struct hl_userptr *userptr)
1320{
1321	struct page **pages;
1322
1323	hl_debugfs_remove_userptr(hdev, userptr);
1324
1325	if (userptr->dma_mapped)
1326		hdev->asic_funcs->hl_dma_unmap_sg(hdev,
1327				userptr->sgt->sgl,
1328				userptr->sgt->nents,
1329				userptr->dir);
1330
1331	pages = frame_vector_pages(userptr->vec);
1332	if (!IS_ERR(pages)) {
1333		int i;
1334
1335		for (i = 0; i < frame_vector_count(userptr->vec); i++)
1336			set_page_dirty_lock(pages[i]);
1337	}
1338	put_vaddr_frames(userptr->vec);
1339	frame_vector_destroy(userptr->vec);
1340
1341	list_del(&userptr->job_node);
1342
1343	sg_free_table(userptr->sgt);
1344	kfree(userptr->sgt);
1345
1346	return 0;
1347}
1348
1349/*
1350 * hl_userptr_delete_list - clear userptr list
1351 *
1352 * @hdev                : pointer to the habanalabs device structure
1353 * @userptr_list        : pointer to the list to clear
1354 *
1355 * This function does the following:
1356 * - Iterates over the list and unpins the host memory and frees the userptr
1357 *   structure.
1358 */
1359void hl_userptr_delete_list(struct hl_device *hdev,
1360				struct list_head *userptr_list)
1361{
1362	struct hl_userptr *userptr, *tmp;
1363
1364	list_for_each_entry_safe(userptr, tmp, userptr_list, job_node) {
1365		hl_unpin_host_memory(hdev, userptr);
1366		kfree(userptr);
1367	}
1368
1369	INIT_LIST_HEAD(userptr_list);
1370}
1371
1372/*
1373 * hl_userptr_is_pinned - returns whether the given userptr is pinned
1374 *
1375 * @hdev                : pointer to the habanalabs device structure
1376 * @userptr_list        : pointer to the list to clear
1377 * @userptr             : pointer to userptr to check
1378 *
1379 * This function does the following:
1380 * - Iterates over the list and checks if the given userptr is in it, means is
1381 *   pinned. If so, returns true, otherwise returns false.
1382 */
1383bool hl_userptr_is_pinned(struct hl_device *hdev, u64 addr,
1384				u32 size, struct list_head *userptr_list,
1385				struct hl_userptr **userptr)
1386{
1387	list_for_each_entry((*userptr), userptr_list, job_node) {
1388		if ((addr == (*userptr)->addr) && (size == (*userptr)->size))
1389			return true;
1390	}
1391
1392	return false;
1393}
1394
1395/*
1396 * hl_va_range_init - initialize virtual addresses range
1397 *
1398 * @hdev                : pointer to the habanalabs device structure
1399 * @va_range            : pointer to the range to initialize
1400 * @start               : range start address
1401 * @end                 : range end address
1402 *
1403 * This function does the following:
1404 * - Initializes the virtual addresses list of the given range with the given
1405 *   addresses.
1406 */
1407static int hl_va_range_init(struct hl_device *hdev,
1408		struct hl_va_range *va_range, u64 start, u64 end)
1409{
1410	int rc;
1411
1412	INIT_LIST_HEAD(&va_range->list);
1413
1414	/* PAGE_SIZE alignment */
1415
1416	if (start & (PAGE_SIZE - 1)) {
1417		start &= PAGE_MASK;
1418		start += PAGE_SIZE;
1419	}
1420
1421	if (end & (PAGE_SIZE - 1))
1422		end &= PAGE_MASK;
1423
1424	if (start >= end) {
1425		dev_err(hdev->dev, "too small vm range for va list\n");
1426		return -EFAULT;
1427	}
1428
1429	rc = add_va_block(hdev, va_range, start, end);
1430
1431	if (rc) {
1432		dev_err(hdev->dev, "Failed to init host va list\n");
1433		return rc;
1434	}
1435
1436	va_range->start_addr = start;
1437	va_range->end_addr = end;
1438
1439	return 0;
1440}
1441
1442/*
1443 * hl_vm_ctx_init_with_ranges - initialize virtual memory for context
1444 *
1445 * @ctx                 : pointer to the habanalabs context structure
1446 * @host_range_start    : host virtual addresses range start
1447 * @host_range_end      : host virtual addresses range end
1448 * @dram_range_start    : dram virtual addresses range start
1449 * @dram_range_end      : dram virtual addresses range end
1450 *
1451 * This function initializes the following:
1452 * - MMU for context
1453 * - Virtual address to area descriptor hashtable
1454 * - Virtual block list of available virtual memory
1455 */
1456static int hl_vm_ctx_init_with_ranges(struct hl_ctx *ctx, u64 host_range_start,
1457				u64 host_range_end, u64 dram_range_start,
1458				u64 dram_range_end)
1459{
1460	struct hl_device *hdev = ctx->hdev;
1461	int rc;
1462
1463	rc = hl_mmu_ctx_init(ctx);
1464	if (rc) {
1465		dev_err(hdev->dev, "failed to init context %d\n", ctx->asid);
1466		return rc;
1467	}
1468
1469	mutex_init(&ctx->mem_hash_lock);
1470	hash_init(ctx->mem_hash);
1471
1472	mutex_init(&ctx->host_va_range.lock);
1473
1474	rc = hl_va_range_init(hdev, &ctx->host_va_range, host_range_start,
1475			host_range_end);
1476	if (rc) {
1477		dev_err(hdev->dev, "failed to init host vm range\n");
1478		goto host_vm_err;
1479	}
1480
1481	mutex_init(&ctx->dram_va_range.lock);
1482
1483	rc = hl_va_range_init(hdev, &ctx->dram_va_range, dram_range_start,
1484			dram_range_end);
1485	if (rc) {
1486		dev_err(hdev->dev, "failed to init dram vm range\n");
1487		goto dram_vm_err;
1488	}
1489
1490	hl_debugfs_add_ctx_mem_hash(hdev, ctx);
1491
1492	return 0;
1493
1494dram_vm_err:
1495	mutex_destroy(&ctx->dram_va_range.lock);
1496
1497	mutex_lock(&ctx->host_va_range.lock);
1498	clear_va_list_locked(hdev, &ctx->host_va_range.list);
1499	mutex_unlock(&ctx->host_va_range.lock);
1500host_vm_err:
1501	mutex_destroy(&ctx->host_va_range.lock);
1502	mutex_destroy(&ctx->mem_hash_lock);
1503	hl_mmu_ctx_fini(ctx);
1504
1505	return rc;
1506}
1507
1508int hl_vm_ctx_init(struct hl_ctx *ctx)
1509{
1510	struct asic_fixed_properties *prop = &ctx->hdev->asic_prop;
1511	u64 host_range_start, host_range_end, dram_range_start,
1512		dram_range_end;
1513
1514	atomic64_set(&ctx->dram_phys_mem, 0);
1515
1516	/*
1517	 * - If MMU is enabled, init the ranges as usual.
1518	 * - If MMU is disabled, in case of host mapping, the returned address
1519	 *   is the given one.
1520	 *   In case of DRAM mapping, the returned address is the physical
1521	 *   address of the memory related to the given handle.
1522	 */
1523	if (ctx->hdev->mmu_enable) {
1524		dram_range_start = prop->va_space_dram_start_address;
1525		dram_range_end = prop->va_space_dram_end_address;
1526		host_range_start = prop->va_space_host_start_address;
1527		host_range_end = prop->va_space_host_end_address;
1528	} else {
1529		dram_range_start = prop->dram_user_base_address;
1530		dram_range_end = prop->dram_end_address;
1531		host_range_start = prop->dram_user_base_address;
1532		host_range_end = prop->dram_end_address;
1533	}
1534
1535	return hl_vm_ctx_init_with_ranges(ctx, host_range_start, host_range_end,
1536			dram_range_start, dram_range_end);
1537}
1538
1539/*
1540 * hl_va_range_fini     - clear a virtual addresses range
1541 *
1542 * @hdev                : pointer to the habanalabs structure
1543 * va_range             : pointer to virtual addresses range
1544 *
1545 * This function initializes the following:
1546 * - Checks that the given range contains the whole initial range
1547 * - Frees the virtual addresses block list and its lock
1548 */
1549static void hl_va_range_fini(struct hl_device *hdev,
1550		struct hl_va_range *va_range)
1551{
1552	struct hl_vm_va_block *va_block;
1553
1554	if (list_empty(&va_range->list)) {
1555		dev_warn(hdev->dev,
1556				"va list should not be empty on cleanup!\n");
1557		goto out;
1558	}
1559
1560	if (!list_is_singular(&va_range->list)) {
1561		dev_warn(hdev->dev,
1562			"va list should not contain multiple blocks on cleanup!\n");
1563		goto free_va_list;
1564	}
1565
1566	va_block = list_first_entry(&va_range->list, typeof(*va_block), node);
1567
1568	if (va_block->start != va_range->start_addr ||
1569		va_block->end != va_range->end_addr) {
1570		dev_warn(hdev->dev,
1571			"wrong va block on cleanup, from 0x%llx to 0x%llx\n",
1572				va_block->start, va_block->end);
1573		goto free_va_list;
1574	}
1575
1576free_va_list:
1577	mutex_lock(&va_range->lock);
1578	clear_va_list_locked(hdev, &va_range->list);
1579	mutex_unlock(&va_range->lock);
1580
1581out:
1582	mutex_destroy(&va_range->lock);
1583}
1584
1585/*
1586 * hl_vm_ctx_fini       - virtual memory teardown of context
1587 *
1588 * @ctx                 : pointer to the habanalabs context structure
1589 *
1590 * This function perform teardown the following:
1591 * - Virtual block list of available virtual memory
1592 * - Virtual address to area descriptor hashtable
1593 * - MMU for context
1594 *
1595 * In addition this function does the following:
1596 * - Unmaps the existing hashtable nodes if the hashtable is not empty. The
1597 *   hashtable should be empty as no valid mappings should exist at this
1598 *   point.
1599 * - Frees any existing physical page list from the idr which relates to the
1600 *   current context asid.
1601 * - This function checks the virtual block list for correctness. At this point
1602 *   the list should contain one element which describes the whole virtual
1603 *   memory range of the context. Otherwise, a warning is printed.
1604 */
1605void hl_vm_ctx_fini(struct hl_ctx *ctx)
1606{
1607	struct hl_device *hdev = ctx->hdev;
1608	struct hl_vm *vm = &hdev->vm;
1609	struct hl_vm_phys_pg_pack *phys_pg_list;
1610	struct hl_vm_hash_node *hnode;
1611	struct hlist_node *tmp_node;
1612	int i;
1613
1614	hl_debugfs_remove_ctx_mem_hash(hdev, ctx);
1615
1616	if (!hash_empty(ctx->mem_hash))
1617		dev_notice(hdev->dev, "ctx is freed while it has va in use\n");
1618
1619	hash_for_each_safe(ctx->mem_hash, i, tmp_node, hnode, node) {
1620		dev_dbg(hdev->dev,
1621			"hl_mem_hash_node of vaddr 0x%llx of asid %d is still alive\n",
1622			hnode->vaddr, ctx->asid);
1623		unmap_device_va(ctx, hnode->vaddr);
1624	}
1625
1626	spin_lock(&vm->idr_lock);
1627	idr_for_each_entry(&vm->phys_pg_pack_handles, phys_pg_list, i)
1628		if (phys_pg_list->asid == ctx->asid) {
1629			dev_dbg(hdev->dev,
1630				"page list 0x%p of asid %d is still alive\n",
1631				phys_pg_list, ctx->asid);
1632			atomic64_sub(phys_pg_list->total_size,
1633					&hdev->dram_used_mem);
1634			free_phys_pg_pack(hdev, phys_pg_list);
1635			idr_remove(&vm->phys_pg_pack_handles, i);
1636		}
1637	spin_unlock(&vm->idr_lock);
1638
1639	hl_va_range_fini(hdev, &ctx->dram_va_range);
1640	hl_va_range_fini(hdev, &ctx->host_va_range);
1641
1642	mutex_destroy(&ctx->mem_hash_lock);
1643	hl_mmu_ctx_fini(ctx);
1644}
1645
1646/*
1647 * hl_vm_init           - initialize virtual memory module
1648 *
1649 * @hdev                : pointer to the habanalabs device structure
1650 *
1651 * This function initializes the following:
1652 * - MMU module
1653 * - DRAM physical pages pool of 2MB
1654 * - Idr for device memory allocation handles
1655 */
1656int hl_vm_init(struct hl_device *hdev)
1657{
1658	struct asic_fixed_properties *prop = &hdev->asic_prop;
1659	struct hl_vm *vm = &hdev->vm;
1660	int rc;
1661
1662	vm->dram_pg_pool = gen_pool_create(__ffs(prop->dram_page_size), -1);
1663	if (!vm->dram_pg_pool) {
1664		dev_err(hdev->dev, "Failed to create dram page pool\n");
1665		return -ENOMEM;
1666	}
1667
1668	kref_init(&vm->dram_pg_pool_refcount);
1669
1670	rc = gen_pool_add(vm->dram_pg_pool, prop->dram_user_base_address,
1671			prop->dram_end_address - prop->dram_user_base_address,
1672			-1);
1673
1674	if (rc) {
1675		dev_err(hdev->dev,
1676			"Failed to add memory to dram page pool %d\n", rc);
1677		goto pool_add_err;
1678	}
1679
1680	spin_lock_init(&vm->idr_lock);
1681	idr_init(&vm->phys_pg_pack_handles);
1682
1683	atomic64_set(&hdev->dram_used_mem, 0);
1684
1685	vm->init_done = true;
1686
1687	return 0;
1688
1689pool_add_err:
1690	gen_pool_destroy(vm->dram_pg_pool);
1691
1692	return rc;
1693}
1694
1695/*
1696 * hl_vm_fini           - virtual memory module teardown
1697 *
1698 * @hdev                : pointer to the habanalabs device structure
1699 *
1700 * This function perform teardown to the following:
1701 * - Idr for device memory allocation handles
1702 * - DRAM physical pages pool of 2MB
1703 * - MMU module
1704 */
1705void hl_vm_fini(struct hl_device *hdev)
1706{
1707	struct hl_vm *vm = &hdev->vm;
1708
1709	if (!vm->init_done)
1710		return;
1711
1712	/*
1713	 * At this point all the contexts should be freed and hence no DRAM
1714	 * memory should be in use. Hence the DRAM pool should be freed here.
1715	 */
1716	if (kref_put(&vm->dram_pg_pool_refcount, dram_pg_pool_do_release) != 1)
1717		dev_warn(hdev->dev, "dram_pg_pool was not destroyed on %s\n",
1718				__func__);
1719
1720	vm->init_done = false;
1721}