Linux Audio

Check our new training course

Loading...
v6.2
   1// SPDX-License-Identifier: MIT
   2/*
   3 * Copyright © 2019 Intel Corporation
   4 */
   5
   6#include <linux/prime_numbers.h>
   7#include <linux/sort.h>
   8
   9#include <drm/drm_buddy.h>
  10
  11#include "../i915_selftest.h"
  12
  13#include "mock_drm.h"
  14#include "mock_gem_device.h"
  15#include "mock_region.h"
  16
  17#include "gem/i915_gem_context.h"
  18#include "gem/i915_gem_lmem.h"
  19#include "gem/i915_gem_region.h"
  20#include "gem/i915_gem_ttm.h"
  21#include "gem/selftests/igt_gem_utils.h"
  22#include "gem/selftests/mock_context.h"
  23#include "gt/intel_engine_pm.h"
  24#include "gt/intel_engine_user.h"
  25#include "gt/intel_gt.h"
  26#include "gt/intel_migrate.h"
  27#include "i915_memcpy.h"
  28#include "i915_ttm_buddy_manager.h"
  29#include "selftests/igt_flush_test.h"
  30#include "selftests/i915_random.h"
  31
  32static void close_objects(struct intel_memory_region *mem,
  33			  struct list_head *objects)
  34{
  35	struct drm_i915_private *i915 = mem->i915;
  36	struct drm_i915_gem_object *obj, *on;
  37
  38	list_for_each_entry_safe(obj, on, objects, st_link) {
  39		i915_gem_object_lock(obj, NULL);
  40		if (i915_gem_object_has_pinned_pages(obj))
  41			i915_gem_object_unpin_pages(obj);
  42		/* No polluting the memory region between tests */
  43		__i915_gem_object_put_pages(obj);
  44		i915_gem_object_unlock(obj);
  45		list_del(&obj->st_link);
  46		i915_gem_object_put(obj);
  47	}
  48
  49	cond_resched();
  50
  51	i915_gem_drain_freed_objects(i915);
  52}
  53
  54static int igt_mock_fill(void *arg)
  55{
  56	struct intel_memory_region *mem = arg;
  57	resource_size_t total = resource_size(&mem->region);
  58	resource_size_t page_size;
  59	resource_size_t rem;
  60	unsigned long max_pages;
  61	unsigned long page_num;
  62	LIST_HEAD(objects);
  63	int err = 0;
  64
  65	page_size = PAGE_SIZE;
  66	max_pages = div64_u64(total, page_size);
  67	rem = total;
  68
  69	for_each_prime_number_from(page_num, 1, max_pages) {
  70		resource_size_t size = page_num * page_size;
  71		struct drm_i915_gem_object *obj;
  72
  73		obj = i915_gem_object_create_region(mem, size, 0, 0);
  74		if (IS_ERR(obj)) {
  75			err = PTR_ERR(obj);
  76			break;
  77		}
  78
  79		err = i915_gem_object_pin_pages_unlocked(obj);
  80		if (err) {
  81			i915_gem_object_put(obj);
  82			break;
  83		}
  84
  85		list_add(&obj->st_link, &objects);
  86		rem -= size;
  87	}
  88
  89	if (err == -ENOMEM)
  90		err = 0;
  91	if (err == -ENXIO) {
  92		if (page_num * page_size <= rem) {
  93			pr_err("%s failed, space still left in region\n",
  94			       __func__);
  95			err = -EINVAL;
  96		} else {
  97			err = 0;
  98		}
  99	}
 100
 101	close_objects(mem, &objects);
 102
 103	return err;
 104}
 105
 106static struct drm_i915_gem_object *
 107igt_object_create(struct intel_memory_region *mem,
 108		  struct list_head *objects,
 109		  u64 size,
 110		  unsigned int flags)
 111{
 112	struct drm_i915_gem_object *obj;
 113	int err;
 114
 115	obj = i915_gem_object_create_region(mem, size, 0, flags);
 116	if (IS_ERR(obj))
 117		return obj;
 118
 119	err = i915_gem_object_pin_pages_unlocked(obj);
 120	if (err)
 121		goto put;
 122
 123	list_add(&obj->st_link, objects);
 124	return obj;
 125
 126put:
 127	i915_gem_object_put(obj);
 128	return ERR_PTR(err);
 129}
 130
 131static void igt_object_release(struct drm_i915_gem_object *obj)
 132{
 133	i915_gem_object_lock(obj, NULL);
 134	i915_gem_object_unpin_pages(obj);
 135	__i915_gem_object_put_pages(obj);
 136	i915_gem_object_unlock(obj);
 137	list_del(&obj->st_link);
 138	i915_gem_object_put(obj);
 139}
 140
 141static bool is_contiguous(struct drm_i915_gem_object *obj)
 142{
 143	struct scatterlist *sg;
 144	dma_addr_t addr = -1;
 145
 146	for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
 147		if (addr != -1 && sg_dma_address(sg) != addr)
 148			return false;
 149
 150		addr = sg_dma_address(sg) + sg_dma_len(sg);
 151	}
 152
 153	return true;
 154}
 155
 156static int igt_mock_reserve(void *arg)
 157{
 158	struct intel_memory_region *mem = arg;
 159	struct drm_i915_private *i915 = mem->i915;
 160	resource_size_t avail = resource_size(&mem->region);
 161	struct drm_i915_gem_object *obj;
 162	const u32 chunk_size = SZ_32M;
 163	u32 i, offset, count, *order;
 164	u64 allocated, cur_avail;
 165	I915_RND_STATE(prng);
 166	LIST_HEAD(objects);
 167	int err = 0;
 168
 169	count = avail / chunk_size;
 170	order = i915_random_order(count, &prng);
 171	if (!order)
 172		return 0;
 173
 174	mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
 175	if (IS_ERR(mem)) {
 176		pr_err("failed to create memory region\n");
 177		err = PTR_ERR(mem);
 178		goto out_free_order;
 179	}
 180
 181	/* Reserve a bunch of ranges within the region */
 182	for (i = 0; i < count; ++i) {
 183		u64 start = order[i] * chunk_size;
 184		u64 size = i915_prandom_u32_max_state(chunk_size, &prng);
 185
 186		/* Allow for some really big holes */
 187		if (!size)
 188			continue;
 189
 190		size = round_up(size, PAGE_SIZE);
 191		offset = igt_random_offset(&prng, 0, chunk_size, size,
 192					   PAGE_SIZE);
 193
 194		err = intel_memory_region_reserve(mem, start + offset, size);
 195		if (err) {
 196			pr_err("%s failed to reserve range", __func__);
 197			goto out_close;
 198		}
 199
 200		/* XXX: maybe sanity check the block range here? */
 201		avail -= size;
 202	}
 203
 204	/* Try to see if we can allocate from the remaining space */
 205	allocated = 0;
 206	cur_avail = avail;
 207	do {
 208		u32 size = i915_prandom_u32_max_state(cur_avail, &prng);
 209
 210		size = max_t(u32, round_up(size, PAGE_SIZE), PAGE_SIZE);
 211		obj = igt_object_create(mem, &objects, size, 0);
 212		if (IS_ERR(obj)) {
 213			if (PTR_ERR(obj) == -ENXIO)
 214				break;
 215
 216			err = PTR_ERR(obj);
 217			goto out_close;
 218		}
 219		cur_avail -= size;
 220		allocated += size;
 221	} while (1);
 222
 223	if (allocated != avail) {
 224		pr_err("%s mismatch between allocation and free space", __func__);
 225		err = -EINVAL;
 226	}
 227
 228out_close:
 229	close_objects(mem, &objects);
 230	intel_memory_region_destroy(mem);
 231out_free_order:
 232	kfree(order);
 233	return err;
 234}
 235
 236static int igt_mock_contiguous(void *arg)
 237{
 238	struct intel_memory_region *mem = arg;
 239	struct drm_i915_gem_object *obj;
 240	unsigned long n_objects;
 241	LIST_HEAD(objects);
 242	LIST_HEAD(holes);
 243	I915_RND_STATE(prng);
 244	resource_size_t total;
 245	resource_size_t min;
 246	u64 target;
 247	int err = 0;
 248
 249	total = resource_size(&mem->region);
 250
 251	/* Min size */
 252	obj = igt_object_create(mem, &objects, PAGE_SIZE,
 253				I915_BO_ALLOC_CONTIGUOUS);
 254	if (IS_ERR(obj))
 255		return PTR_ERR(obj);
 256
 257	if (!is_contiguous(obj)) {
 258		pr_err("%s min object spans disjoint sg entries\n", __func__);
 259		err = -EINVAL;
 260		goto err_close_objects;
 261	}
 262
 263	igt_object_release(obj);
 264
 265	/* Max size */
 266	obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
 267	if (IS_ERR(obj))
 268		return PTR_ERR(obj);
 269
 270	if (!is_contiguous(obj)) {
 271		pr_err("%s max object spans disjoint sg entries\n", __func__);
 272		err = -EINVAL;
 273		goto err_close_objects;
 274	}
 275
 276	igt_object_release(obj);
 277
 278	/* Internal fragmentation should not bleed into the object size */
 279	target = i915_prandom_u64_state(&prng);
 280	div64_u64_rem(target, total, &target);
 281	target = round_up(target, PAGE_SIZE);
 282	target = max_t(u64, PAGE_SIZE, target);
 283
 284	obj = igt_object_create(mem, &objects, target,
 285				I915_BO_ALLOC_CONTIGUOUS);
 286	if (IS_ERR(obj))
 287		return PTR_ERR(obj);
 288
 289	if (obj->base.size != target) {
 290		pr_err("%s obj->base.size(%zx) != target(%llx)\n", __func__,
 291		       obj->base.size, target);
 292		err = -EINVAL;
 293		goto err_close_objects;
 294	}
 295
 296	if (!is_contiguous(obj)) {
 297		pr_err("%s object spans disjoint sg entries\n", __func__);
 298		err = -EINVAL;
 299		goto err_close_objects;
 300	}
 301
 302	igt_object_release(obj);
 303
 304	/*
 305	 * Try to fragment the address space, such that half of it is free, but
 306	 * the max contiguous block size is SZ_64K.
 307	 */
 308
 309	target = SZ_64K;
 310	n_objects = div64_u64(total, target);
 311
 312	while (n_objects--) {
 313		struct list_head *list;
 314
 315		if (n_objects % 2)
 316			list = &holes;
 317		else
 318			list = &objects;
 319
 320		obj = igt_object_create(mem, list, target,
 321					I915_BO_ALLOC_CONTIGUOUS);
 322		if (IS_ERR(obj)) {
 323			err = PTR_ERR(obj);
 324			goto err_close_objects;
 325		}
 326	}
 327
 328	close_objects(mem, &holes);
 329
 330	min = target;
 331	target = total >> 1;
 332
 333	/* Make sure we can still allocate all the fragmented space */
 334	obj = igt_object_create(mem, &objects, target, 0);
 335	if (IS_ERR(obj)) {
 336		err = PTR_ERR(obj);
 337		goto err_close_objects;
 338	}
 339
 340	igt_object_release(obj);
 341
 342	/*
 343	 * Even though we have enough free space, we don't have a big enough
 344	 * contiguous block. Make sure that holds true.
 345	 */
 346
 347	do {
 348		bool should_fail = target > min;
 349
 350		obj = igt_object_create(mem, &objects, target,
 351					I915_BO_ALLOC_CONTIGUOUS);
 352		if (should_fail != IS_ERR(obj)) {
 353			pr_err("%s target allocation(%llx) mismatch\n",
 354			       __func__, target);
 355			err = -EINVAL;
 356			goto err_close_objects;
 357		}
 358
 359		target >>= 1;
 360	} while (target >= PAGE_SIZE);
 361
 362err_close_objects:
 363	list_splice_tail(&holes, &objects);
 364	close_objects(mem, &objects);
 365	return err;
 366}
 367
 368static int igt_mock_splintered_region(void *arg)
 369{
 370	struct intel_memory_region *mem = arg;
 371	struct drm_i915_private *i915 = mem->i915;
 372	struct i915_ttm_buddy_resource *res;
 373	struct drm_i915_gem_object *obj;
 374	struct drm_buddy *mm;
 375	unsigned int expected_order;
 376	LIST_HEAD(objects);
 377	u64 size;
 378	int err = 0;
 379
 380	/*
 381	 * Sanity check we can still allocate everything even if the
 382	 * mm.max_order != mm.size. i.e our starting address space size is not a
 383	 * power-of-two.
 384	 */
 385
 386	size = (SZ_4G - 1) & PAGE_MASK;
 387	mem = mock_region_create(i915, 0, size, PAGE_SIZE, 0, 0);
 388	if (IS_ERR(mem))
 389		return PTR_ERR(mem);
 390
 391	obj = igt_object_create(mem, &objects, size, 0);
 392	if (IS_ERR(obj)) {
 393		err = PTR_ERR(obj);
 394		goto out_close;
 395	}
 396
 397	res = to_ttm_buddy_resource(obj->mm.res);
 398	mm = res->mm;
 399	if (mm->size != size) {
 400		pr_err("%s size mismatch(%llu != %llu)\n",
 401		       __func__, mm->size, size);
 402		err = -EINVAL;
 403		goto out_put;
 404	}
 405
 406	expected_order = get_order(rounddown_pow_of_two(size));
 407	if (mm->max_order != expected_order) {
 408		pr_err("%s order mismatch(%u != %u)\n",
 409		       __func__, mm->max_order, expected_order);
 410		err = -EINVAL;
 411		goto out_put;
 412	}
 413
 414	close_objects(mem, &objects);
 415
 416	/*
 417	 * While we should be able allocate everything without any flag
 418	 * restrictions, if we consider I915_BO_ALLOC_CONTIGUOUS then we are
 419	 * actually limited to the largest power-of-two for the region size i.e
 420	 * max_order, due to the inner workings of the buddy allocator. So make
 421	 * sure that does indeed hold true.
 422	 */
 423
 424	obj = igt_object_create(mem, &objects, size, I915_BO_ALLOC_CONTIGUOUS);
 425	if (!IS_ERR(obj)) {
 426		pr_err("%s too large contiguous allocation was not rejected\n",
 427		       __func__);
 428		err = -EINVAL;
 429		goto out_close;
 430	}
 431
 432	obj = igt_object_create(mem, &objects, rounddown_pow_of_two(size),
 433				I915_BO_ALLOC_CONTIGUOUS);
 434	if (IS_ERR(obj)) {
 435		pr_err("%s largest possible contiguous allocation failed\n",
 436		       __func__);
 437		err = PTR_ERR(obj);
 438		goto out_close;
 439	}
 440
 441out_close:
 442	close_objects(mem, &objects);
 443out_put:
 444	intel_memory_region_destroy(mem);
 445	return err;
 446}
 447
 448#ifndef SZ_8G
 449#define SZ_8G BIT_ULL(33)
 450#endif
 451
 452static int igt_mock_max_segment(void *arg)
 453{
 454	struct intel_memory_region *mem = arg;
 455	struct drm_i915_private *i915 = mem->i915;
 456	struct i915_ttm_buddy_resource *res;
 457	struct drm_i915_gem_object *obj;
 458	struct drm_buddy_block *block;
 459	struct drm_buddy *mm;
 460	struct list_head *blocks;
 461	struct scatterlist *sg;
 462	I915_RND_STATE(prng);
 463	LIST_HEAD(objects);
 464	unsigned int max_segment;
 465	unsigned int ps;
 466	u64 size;
 467	int err = 0;
 468
 469	/*
 470	 * While we may create very large contiguous blocks, we may need
 471	 * to break those down for consumption elsewhere. In particular,
 472	 * dma-mapping with scatterlist elements have an implicit limit of
 473	 * UINT_MAX on each element.
 474	 */
 475
 476	size = SZ_8G;
 477	ps = PAGE_SIZE;
 478	if (i915_prandom_u64_state(&prng) & 1)
 479		ps = SZ_64K; /* For something like DG2 */
 480
 481	max_segment = round_down(UINT_MAX, ps);
 482
 483	mem = mock_region_create(i915, 0, size, ps, 0, 0);
 484	if (IS_ERR(mem))
 485		return PTR_ERR(mem);
 486
 487	obj = igt_object_create(mem, &objects, size, 0);
 488	if (IS_ERR(obj)) {
 489		err = PTR_ERR(obj);
 490		goto out_put;
 491	}
 492
 493	res = to_ttm_buddy_resource(obj->mm.res);
 494	blocks = &res->blocks;
 495	mm = res->mm;
 496	size = 0;
 497	list_for_each_entry(block, blocks, link) {
 498		if (drm_buddy_block_size(mm, block) > size)
 499			size = drm_buddy_block_size(mm, block);
 500	}
 501	if (size < max_segment) {
 502		pr_err("%s: Failed to create a huge contiguous block [> %u], largest block %lld\n",
 503		       __func__, max_segment, size);
 504		err = -EINVAL;
 505		goto out_close;
 506	}
 507
 508	for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
 509		dma_addr_t daddr = sg_dma_address(sg);
 510
 511		if (sg->length > max_segment) {
 512			pr_err("%s: Created an oversized scatterlist entry, %u > %u\n",
 513			       __func__, sg->length, max_segment);
 514			err = -EINVAL;
 515			goto out_close;
 516		}
 517
 518		if (!IS_ALIGNED(daddr, ps)) {
 519			pr_err("%s: Created an unaligned scatterlist entry, addr=%pa, ps=%u\n",
 520			       __func__,  &daddr, ps);
 521			err = -EINVAL;
 522			goto out_close;
 523		}
 524	}
 525
 526out_close:
 527	close_objects(mem, &objects);
 528out_put:
 529	intel_memory_region_destroy(mem);
 530	return err;
 531}
 532
 533static u64 igt_object_mappable_total(struct drm_i915_gem_object *obj)
 534{
 535	struct intel_memory_region *mr = obj->mm.region;
 536	struct i915_ttm_buddy_resource *bman_res =
 537		to_ttm_buddy_resource(obj->mm.res);
 538	struct drm_buddy *mm = bman_res->mm;
 539	struct drm_buddy_block *block;
 540	u64 total;
 541
 542	total = 0;
 543	list_for_each_entry(block, &bman_res->blocks, link) {
 544		u64 start = drm_buddy_block_offset(block);
 545		u64 end = start + drm_buddy_block_size(mm, block);
 546
 547		if (start < mr->io_size)
 548			total += min_t(u64, end, mr->io_size) - start;
 549	}
 550
 551	return total;
 552}
 553
 554static int igt_mock_io_size(void *arg)
 555{
 556	struct intel_memory_region *mr = arg;
 557	struct drm_i915_private *i915 = mr->i915;
 558	struct drm_i915_gem_object *obj;
 559	u64 mappable_theft_total;
 560	u64 io_size;
 561	u64 total;
 562	u64 ps;
 563	u64 rem;
 564	u64 size;
 565	I915_RND_STATE(prng);
 566	LIST_HEAD(objects);
 567	int err = 0;
 568
 569	ps = SZ_4K;
 570	if (i915_prandom_u64_state(&prng) & 1)
 571		ps = SZ_64K; /* For something like DG2 */
 572
 573	div64_u64_rem(i915_prandom_u64_state(&prng), SZ_8G, &total);
 574	total = round_down(total, ps);
 575	total = max_t(u64, total, SZ_1G);
 576
 577	div64_u64_rem(i915_prandom_u64_state(&prng), total - ps, &io_size);
 578	io_size = round_down(io_size, ps);
 579	io_size = max_t(u64, io_size, SZ_256M); /* 256M seems to be the common lower limit */
 580
 581	pr_info("%s with ps=%llx, io_size=%llx, total=%llx\n",
 582		__func__, ps, io_size, total);
 583
 584	mr = mock_region_create(i915, 0, total, ps, 0, io_size);
 585	if (IS_ERR(mr)) {
 586		err = PTR_ERR(mr);
 587		goto out_err;
 588	}
 589
 590	mappable_theft_total = 0;
 591	rem = total - io_size;
 592	do {
 593		div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
 594		size = round_down(size, ps);
 595		size = max(size, ps);
 596
 597		obj = igt_object_create(mr, &objects, size,
 598					I915_BO_ALLOC_GPU_ONLY);
 599		if (IS_ERR(obj)) {
 600			pr_err("%s TOPDOWN failed with rem=%llx, size=%llx\n",
 601			       __func__, rem, size);
 602			err = PTR_ERR(obj);
 603			goto out_close;
 604		}
 605
 606		mappable_theft_total += igt_object_mappable_total(obj);
 607		rem -= size;
 608	} while (rem);
 609
 610	pr_info("%s mappable theft=(%lluMiB/%lluMiB), total=%lluMiB\n",
 611		__func__,
 612		(u64)mappable_theft_total >> 20,
 613		(u64)io_size >> 20,
 614		(u64)total >> 20);
 615
 616	/*
 617	 * Even if we allocate all of the non-mappable portion, we should still
 618	 * be able to dip into the mappable portion.
 619	 */
 620	obj = igt_object_create(mr, &objects, io_size,
 621				I915_BO_ALLOC_GPU_ONLY);
 622	if (IS_ERR(obj)) {
 623		pr_err("%s allocation unexpectedly failed\n", __func__);
 624		err = PTR_ERR(obj);
 625		goto out_close;
 626	}
 627
 628	close_objects(mr, &objects);
 629
 630	rem = io_size;
 631	do {
 632		div64_u64_rem(i915_prandom_u64_state(&prng), rem, &size);
 633		size = round_down(size, ps);
 634		size = max(size, ps);
 635
 636		obj = igt_object_create(mr, &objects, size, 0);
 637		if (IS_ERR(obj)) {
 638			pr_err("%s MAPPABLE failed with rem=%llx, size=%llx\n",
 639			       __func__, rem, size);
 640			err = PTR_ERR(obj);
 641			goto out_close;
 642		}
 643
 644		if (igt_object_mappable_total(obj) != size) {
 645			pr_err("%s allocation is not mappable(size=%llx)\n",
 646			       __func__, size);
 647			err = -EINVAL;
 648			goto out_close;
 649		}
 650		rem -= size;
 651	} while (rem);
 652
 653	/*
 654	 * We assume CPU access is required by default, which should result in a
 655	 * failure here, even though the non-mappable portion is free.
 656	 */
 657	obj = igt_object_create(mr, &objects, ps, 0);
 658	if (!IS_ERR(obj)) {
 659		pr_err("%s allocation unexpectedly succeeded\n", __func__);
 660		err = -EINVAL;
 661		goto out_close;
 662	}
 663
 664out_close:
 665	close_objects(mr, &objects);
 666	intel_memory_region_destroy(mr);
 667out_err:
 668	if (err == -ENOMEM)
 669		err = 0;
 670
 671	return err;
 672}
 673
 674static int igt_gpu_write_dw(struct intel_context *ce,
 675			    struct i915_vma *vma,
 676			    u32 dword,
 677			    u32 value)
 678{
 679	return igt_gpu_fill_dw(ce, vma, dword * sizeof(u32),
 680			       vma->size >> PAGE_SHIFT, value);
 681}
 682
 683static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
 684{
 685	unsigned long n = obj->base.size >> PAGE_SHIFT;
 686	u32 *ptr;
 687	int err;
 688
 689	err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
 690	if (err)
 691		return err;
 692
 693	ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
 694	if (IS_ERR(ptr))
 695		return PTR_ERR(ptr);
 696
 697	ptr += dword;
 698	while (n--) {
 699		if (*ptr != val) {
 700			pr_err("base[%u]=%08x, val=%08x\n",
 701			       dword, *ptr, val);
 702			err = -EINVAL;
 703			break;
 704		}
 705
 706		ptr += PAGE_SIZE / sizeof(*ptr);
 707	}
 708
 709	i915_gem_object_unpin_map(obj);
 710	return err;
 711}
 712
 713static int igt_gpu_write(struct i915_gem_context *ctx,
 714			 struct drm_i915_gem_object *obj)
 715{
 716	struct i915_gem_engines *engines;
 717	struct i915_gem_engines_iter it;
 718	struct i915_address_space *vm;
 719	struct intel_context *ce;
 720	I915_RND_STATE(prng);
 721	IGT_TIMEOUT(end_time);
 722	unsigned int count;
 723	struct i915_vma *vma;
 724	int *order;
 725	int i, n;
 726	int err = 0;
 727
 728	GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
 729
 730	n = 0;
 731	count = 0;
 732	for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
 733		count++;
 734		if (!intel_engine_can_store_dword(ce->engine))
 735			continue;
 736
 737		vm = ce->vm;
 738		n++;
 739	}
 740	i915_gem_context_unlock_engines(ctx);
 741	if (!n)
 742		return 0;
 743
 744	order = i915_random_order(count * count, &prng);
 745	if (!order)
 746		return -ENOMEM;
 747
 748	vma = i915_vma_instance(obj, vm, NULL);
 749	if (IS_ERR(vma)) {
 750		err = PTR_ERR(vma);
 751		goto out_free;
 752	}
 753
 754	err = i915_vma_pin(vma, 0, 0, PIN_USER);
 755	if (err)
 756		goto out_free;
 757
 758	i = 0;
 759	engines = i915_gem_context_lock_engines(ctx);
 760	do {
 761		u32 rng = prandom_u32_state(&prng);
 762		u32 dword = offset_in_page(rng) / 4;
 763
 764		ce = engines->engines[order[i] % engines->num_engines];
 765		i = (i + 1) % (count * count);
 766		if (!ce || !intel_engine_can_store_dword(ce->engine))
 767			continue;
 768
 769		err = igt_gpu_write_dw(ce, vma, dword, rng);
 770		if (err)
 771			break;
 772
 773		i915_gem_object_lock(obj, NULL);
 774		err = igt_cpu_check(obj, dword, rng);
 775		i915_gem_object_unlock(obj);
 776		if (err)
 777			break;
 778	} while (!__igt_timeout(end_time, NULL));
 779	i915_gem_context_unlock_engines(ctx);
 780
 781out_free:
 782	kfree(order);
 783
 784	if (err == -ENOMEM)
 785		err = 0;
 786
 787	return err;
 788}
 789
 790static int igt_lmem_create(void *arg)
 791{
 792	struct drm_i915_private *i915 = arg;
 793	struct drm_i915_gem_object *obj;
 794	int err = 0;
 795
 796	obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
 797	if (IS_ERR(obj))
 798		return PTR_ERR(obj);
 799
 800	err = i915_gem_object_pin_pages_unlocked(obj);
 801	if (err)
 802		goto out_put;
 803
 804	i915_gem_object_unpin_pages(obj);
 805out_put:
 806	i915_gem_object_put(obj);
 807
 808	return err;
 809}
 810
 811static int igt_lmem_create_with_ps(void *arg)
 812{
 813	struct drm_i915_private *i915 = arg;
 814	int err = 0;
 815	u32 ps;
 816
 817	for (ps = PAGE_SIZE; ps <= SZ_1G; ps <<= 1) {
 818		struct drm_i915_gem_object *obj;
 819		dma_addr_t daddr;
 820
 821		obj = __i915_gem_object_create_lmem_with_ps(i915, ps, ps, 0);
 822		if (IS_ERR(obj)) {
 823			err = PTR_ERR(obj);
 824			if (err == -ENXIO || err == -E2BIG) {
 825				pr_info("%s not enough lmem for ps(%u) err=%d\n",
 826					__func__, ps, err);
 827				err = 0;
 828			}
 829
 830			break;
 831		}
 832
 833		if (obj->base.size != ps) {
 834			pr_err("%s size(%zu) != ps(%u)\n",
 835			       __func__, obj->base.size, ps);
 836			err = -EINVAL;
 837			goto out_put;
 838		}
 839
 840		i915_gem_object_lock(obj, NULL);
 841		err = i915_gem_object_pin_pages(obj);
 842		if (err) {
 843			if (err == -ENXIO || err == -E2BIG || err == -ENOMEM) {
 844				pr_info("%s not enough lmem for ps(%u) err=%d\n",
 845					__func__, ps, err);
 846				err = 0;
 847			}
 848			goto out_put;
 849		}
 850
 851		daddr = i915_gem_object_get_dma_address(obj, 0);
 852		if (!IS_ALIGNED(daddr, ps)) {
 853			pr_err("%s daddr(%pa) not aligned with ps(%u)\n",
 854			       __func__, &daddr, ps);
 855			err = -EINVAL;
 856			goto out_unpin;
 857		}
 858
 859out_unpin:
 860		i915_gem_object_unpin_pages(obj);
 861		__i915_gem_object_put_pages(obj);
 862out_put:
 863		i915_gem_object_unlock(obj);
 864		i915_gem_object_put(obj);
 865
 866		if (err)
 867			break;
 868	}
 869
 870	return err;
 871}
 872
 873static int igt_lmem_create_cleared_cpu(void *arg)
 874{
 875	struct drm_i915_private *i915 = arg;
 876	I915_RND_STATE(prng);
 877	IGT_TIMEOUT(end_time);
 878	u32 size, i;
 879	int err;
 880
 881	i915_gem_drain_freed_objects(i915);
 882
 883	size = max_t(u32, PAGE_SIZE, i915_prandom_u32_max_state(SZ_32M, &prng));
 884	size = round_up(size, PAGE_SIZE);
 885	i = 0;
 886
 887	do {
 888		struct drm_i915_gem_object *obj;
 889		unsigned int flags;
 890		u32 dword, val;
 891		void *vaddr;
 892
 893		/*
 894		 * Alternate between cleared and uncleared allocations, while
 895		 * also dirtying the pages each time to check that the pages are
 896		 * always cleared if requested, since we should get some overlap
 897		 * of the underlying pages, if not all, since we are the only
 898		 * user.
 899		 */
 900
 901		flags = I915_BO_ALLOC_CPU_CLEAR;
 902		if (i & 1)
 903			flags = 0;
 904
 905		obj = i915_gem_object_create_lmem(i915, size, flags);
 906		if (IS_ERR(obj))
 907			return PTR_ERR(obj);
 908
 909		i915_gem_object_lock(obj, NULL);
 910		err = i915_gem_object_pin_pages(obj);
 911		if (err)
 912			goto out_put;
 913
 914		dword = i915_prandom_u32_max_state(PAGE_SIZE / sizeof(u32),
 915						   &prng);
 916
 917		if (flags & I915_BO_ALLOC_CPU_CLEAR) {
 918			err = igt_cpu_check(obj, dword, 0);
 919			if (err) {
 920				pr_err("%s failed with size=%u, flags=%u\n",
 921				       __func__, size, flags);
 922				goto out_unpin;
 923			}
 924		}
 925
 926		vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
 927		if (IS_ERR(vaddr)) {
 928			err = PTR_ERR(vaddr);
 929			goto out_unpin;
 930		}
 931
 932		val = prandom_u32_state(&prng);
 933
 934		memset32(vaddr, val, obj->base.size / sizeof(u32));
 935
 936		i915_gem_object_flush_map(obj);
 937		i915_gem_object_unpin_map(obj);
 938out_unpin:
 939		i915_gem_object_unpin_pages(obj);
 940		__i915_gem_object_put_pages(obj);
 941out_put:
 942		i915_gem_object_unlock(obj);
 943		i915_gem_object_put(obj);
 944
 945		if (err)
 946			break;
 947		++i;
 948	} while (!__igt_timeout(end_time, NULL));
 949
 950	pr_info("%s completed (%u) iterations\n", __func__, i);
 951
 952	return err;
 953}
 954
 955static int igt_lmem_write_gpu(void *arg)
 956{
 957	struct drm_i915_private *i915 = arg;
 958	struct drm_i915_gem_object *obj;
 959	struct i915_gem_context *ctx;
 960	struct file *file;
 961	I915_RND_STATE(prng);
 962	u32 sz;
 963	int err;
 964
 965	file = mock_file(i915);
 966	if (IS_ERR(file))
 967		return PTR_ERR(file);
 968
 969	ctx = live_context(i915, file);
 970	if (IS_ERR(ctx)) {
 971		err = PTR_ERR(ctx);
 972		goto out_file;
 973	}
 974
 975	sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
 976
 977	obj = i915_gem_object_create_lmem(i915, sz, 0);
 978	if (IS_ERR(obj)) {
 979		err = PTR_ERR(obj);
 980		goto out_file;
 981	}
 982
 983	err = i915_gem_object_pin_pages_unlocked(obj);
 984	if (err)
 985		goto out_put;
 986
 987	err = igt_gpu_write(ctx, obj);
 988	if (err)
 989		pr_err("igt_gpu_write failed(%d)\n", err);
 990
 991	i915_gem_object_unpin_pages(obj);
 992out_put:
 993	i915_gem_object_put(obj);
 994out_file:
 995	fput(file);
 996	return err;
 997}
 998
 999static struct intel_engine_cs *
1000random_engine_class(struct drm_i915_private *i915,
1001		    unsigned int class,
1002		    struct rnd_state *prng)
1003{
1004	struct intel_engine_cs *engine;
1005	unsigned int count;
1006
1007	count = 0;
1008	for (engine = intel_engine_lookup_user(i915, class, 0);
1009	     engine && engine->uabi_class == class;
1010	     engine = rb_entry_safe(rb_next(&engine->uabi_node),
1011				    typeof(*engine), uabi_node))
1012		count++;
1013
1014	count = i915_prandom_u32_max_state(count, prng);
1015	return intel_engine_lookup_user(i915, class, count);
1016}
1017
1018static int igt_lmem_write_cpu(void *arg)
1019{
1020	struct drm_i915_private *i915 = arg;
1021	struct drm_i915_gem_object *obj;
1022	I915_RND_STATE(prng);
1023	IGT_TIMEOUT(end_time);
1024	u32 bytes[] = {
1025		0, /* rng placeholder */
1026		sizeof(u32),
1027		sizeof(u64),
1028		64, /* cl */
1029		PAGE_SIZE,
1030		PAGE_SIZE - sizeof(u32),
1031		PAGE_SIZE - sizeof(u64),
1032		PAGE_SIZE - 64,
1033	};
1034	struct intel_engine_cs *engine;
1035	struct i915_request *rq;
1036	u32 *vaddr;
1037	u32 sz;
1038	u32 i;
1039	int *order;
1040	int count;
1041	int err;
1042
1043	engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
1044	if (!engine)
1045		return 0;
1046
1047	pr_info("%s: using %s\n", __func__, engine->name);
1048
1049	sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
1050	sz = max_t(u32, 2 * PAGE_SIZE, sz);
1051
1052	obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
1053	if (IS_ERR(obj))
1054		return PTR_ERR(obj);
1055
1056	vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
1057	if (IS_ERR(vaddr)) {
1058		err = PTR_ERR(vaddr);
1059		goto out_put;
1060	}
1061
1062	i915_gem_object_lock(obj, NULL);
1063
1064	err = dma_resv_reserve_fences(obj->base.resv, 1);
1065	if (err) {
1066		i915_gem_object_unlock(obj);
1067		goto out_put;
1068	}
1069
1070	/* Put the pages into a known state -- from the gpu for added fun */
1071	intel_engine_pm_get(engine);
1072	err = intel_context_migrate_clear(engine->gt->migrate.context, NULL,
1073					  obj->mm.pages->sgl, I915_CACHE_NONE,
1074					  true, 0xdeadbeaf, &rq);
1075	if (rq) {
1076		dma_resv_add_fence(obj->base.resv, &rq->fence,
1077				   DMA_RESV_USAGE_WRITE);
1078		i915_request_put(rq);
1079	}
1080
1081	intel_engine_pm_put(engine);
1082	if (!err)
1083		err = i915_gem_object_set_to_wc_domain(obj, true);
 
 
 
1084	i915_gem_object_unlock(obj);
1085	if (err)
1086		goto out_unpin;
1087
1088	count = ARRAY_SIZE(bytes);
1089	order = i915_random_order(count * count, &prng);
1090	if (!order) {
1091		err = -ENOMEM;
1092		goto out_unpin;
1093	}
1094
1095	/* A random multiple of u32, picked between [64, PAGE_SIZE - 64] */
1096	bytes[0] = igt_random_offset(&prng, 64, PAGE_SIZE - 64, 0, sizeof(u32));
1097	GEM_BUG_ON(!IS_ALIGNED(bytes[0], sizeof(u32)));
1098
1099	i = 0;
1100	do {
1101		u32 offset;
1102		u32 align;
1103		u32 dword;
1104		u32 size;
1105		u32 val;
1106
1107		size = bytes[order[i] % count];
1108		i = (i + 1) % (count * count);
1109
1110		align = bytes[order[i] % count];
1111		i = (i + 1) % (count * count);
1112
1113		align = max_t(u32, sizeof(u32), rounddown_pow_of_two(align));
1114
1115		offset = igt_random_offset(&prng, 0, obj->base.size,
1116					   size, align);
1117
1118		val = prandom_u32_state(&prng);
1119		memset32(vaddr + offset / sizeof(u32), val ^ 0xdeadbeaf,
1120			 size / sizeof(u32));
1121
1122		/*
1123		 * Sample random dw -- don't waste precious time reading every
1124		 * single dw.
1125		 */
1126		dword = igt_random_offset(&prng, offset,
1127					  offset + size,
1128					  sizeof(u32), sizeof(u32));
1129		dword /= sizeof(u32);
1130		if (vaddr[dword] != (val ^ 0xdeadbeaf)) {
1131			pr_err("%s vaddr[%u]=%u, val=%u, size=%u, align=%u, offset=%u\n",
1132			       __func__, dword, vaddr[dword], val ^ 0xdeadbeaf,
1133			       size, align, offset);
1134			err = -EINVAL;
1135			break;
1136		}
1137	} while (!__igt_timeout(end_time, NULL));
1138
1139out_unpin:
1140	i915_gem_object_unpin_map(obj);
1141out_put:
1142	i915_gem_object_put(obj);
1143
1144	return err;
1145}
1146
1147static const char *repr_type(u32 type)
1148{
1149	switch (type) {
1150	case I915_MAP_WB:
1151		return "WB";
1152	case I915_MAP_WC:
1153		return "WC";
1154	}
1155
1156	return "";
1157}
1158
1159static struct drm_i915_gem_object *
1160create_region_for_mapping(struct intel_memory_region *mr, u64 size, u32 type,
1161			  void **out_addr)
1162{
1163	struct drm_i915_gem_object *obj;
1164	void *addr;
1165
1166	obj = i915_gem_object_create_region(mr, size, 0, 0);
1167	if (IS_ERR(obj)) {
1168		if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
1169			return ERR_PTR(-ENODEV);
1170		return obj;
1171	}
1172
1173	addr = i915_gem_object_pin_map_unlocked(obj, type);
1174	if (IS_ERR(addr)) {
1175		i915_gem_object_put(obj);
1176		if (PTR_ERR(addr) == -ENXIO)
1177			return ERR_PTR(-ENODEV);
1178		return addr;
1179	}
1180
1181	*out_addr = addr;
1182	return obj;
1183}
1184
1185static int wrap_ktime_compare(const void *A, const void *B)
1186{
1187	const ktime_t *a = A, *b = B;
1188
1189	return ktime_compare(*a, *b);
1190}
1191
1192static void igt_memcpy_long(void *dst, const void *src, size_t size)
1193{
1194	unsigned long *tmp = dst;
1195	const unsigned long *s = src;
1196
1197	size = size / sizeof(unsigned long);
1198	while (size--)
1199		*tmp++ = *s++;
1200}
1201
1202static inline void igt_memcpy(void *dst, const void *src, size_t size)
1203{
1204	memcpy(dst, src, size);
1205}
1206
1207static inline void igt_memcpy_from_wc(void *dst, const void *src, size_t size)
1208{
1209	i915_memcpy_from_wc(dst, src, size);
1210}
1211
1212static int _perf_memcpy(struct intel_memory_region *src_mr,
1213			struct intel_memory_region *dst_mr,
1214			u64 size, u32 src_type, u32 dst_type)
1215{
1216	struct drm_i915_private *i915 = src_mr->i915;
1217	const struct {
1218		const char *name;
1219		void (*copy)(void *dst, const void *src, size_t size);
1220		bool skip;
1221	} tests[] = {
1222		{
1223			"memcpy",
1224			igt_memcpy,
1225		},
1226		{
1227			"memcpy_long",
1228			igt_memcpy_long,
1229		},
1230		{
1231			"memcpy_from_wc",
1232			igt_memcpy_from_wc,
1233			!i915_has_memcpy_from_wc(),
1234		},
1235	};
1236	struct drm_i915_gem_object *src, *dst;
1237	void *src_addr, *dst_addr;
1238	int ret = 0;
1239	int i;
1240
1241	src = create_region_for_mapping(src_mr, size, src_type, &src_addr);
1242	if (IS_ERR(src)) {
1243		ret = PTR_ERR(src);
1244		goto out;
1245	}
1246
1247	dst = create_region_for_mapping(dst_mr, size, dst_type, &dst_addr);
1248	if (IS_ERR(dst)) {
1249		ret = PTR_ERR(dst);
1250		goto out_unpin_src;
1251	}
1252
1253	for (i = 0; i < ARRAY_SIZE(tests); ++i) {
1254		ktime_t t[5];
1255		int pass;
1256
1257		if (tests[i].skip)
1258			continue;
1259
1260		for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
1261			ktime_t t0, t1;
1262
1263			t0 = ktime_get();
1264
1265			tests[i].copy(dst_addr, src_addr, size);
1266
1267			t1 = ktime_get();
1268			t[pass] = ktime_sub(t1, t0);
1269		}
1270
1271		sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
1272		if (t[0] <= 0) {
1273			/* ignore the impossible to protect our sanity */
1274			pr_debug("Skipping %s src(%s, %s) -> dst(%s, %s) %14s %4lluKiB copy, unstable measurement [%lld, %lld]\n",
1275				 __func__,
1276				 src_mr->name, repr_type(src_type),
1277				 dst_mr->name, repr_type(dst_type),
1278				 tests[i].name, size >> 10,
1279				 t[0], t[4]);
1280			continue;
1281		}
1282
1283		pr_info("%s src(%s, %s) -> dst(%s, %s) %14s %4llu KiB copy: %5lld MiB/s\n",
1284			__func__,
1285			src_mr->name, repr_type(src_type),
1286			dst_mr->name, repr_type(dst_type),
1287			tests[i].name, size >> 10,
 
 
 
1288			div64_u64(mul_u32_u32(4 * size,
1289					      1000 * 1000 * 1000),
1290				  t[1] + 2 * t[2] + t[3]) >> 20);
1291
1292		cond_resched();
1293	}
1294
1295	i915_gem_object_unpin_map(dst);
1296	i915_gem_object_put(dst);
1297out_unpin_src:
1298	i915_gem_object_unpin_map(src);
1299	i915_gem_object_put(src);
1300
1301	i915_gem_drain_freed_objects(i915);
1302out:
1303	if (ret == -ENODEV)
1304		ret = 0;
1305
1306	return ret;
1307}
1308
1309static int perf_memcpy(void *arg)
1310{
1311	struct drm_i915_private *i915 = arg;
1312	static const u32 types[] = {
1313		I915_MAP_WB,
1314		I915_MAP_WC,
1315	};
1316	static const u32 sizes[] = {
1317		SZ_4K,
1318		SZ_64K,
1319		SZ_4M,
1320	};
1321	struct intel_memory_region *src_mr, *dst_mr;
1322	int src_id, dst_id;
1323	int i, j, k;
1324	int ret;
1325
1326	for_each_memory_region(src_mr, i915, src_id) {
1327		for_each_memory_region(dst_mr, i915, dst_id) {
1328			for (i = 0; i < ARRAY_SIZE(sizes); ++i) {
1329				for (j = 0; j < ARRAY_SIZE(types); ++j) {
1330					for (k = 0; k < ARRAY_SIZE(types); ++k) {
1331						ret = _perf_memcpy(src_mr,
1332								   dst_mr,
1333								   sizes[i],
1334								   types[j],
1335								   types[k]);
1336						if (ret)
1337							return ret;
1338					}
1339				}
1340			}
1341		}
1342	}
1343
1344	return 0;
1345}
1346
1347int intel_memory_region_mock_selftests(void)
1348{
1349	static const struct i915_subtest tests[] = {
1350		SUBTEST(igt_mock_reserve),
1351		SUBTEST(igt_mock_fill),
1352		SUBTEST(igt_mock_contiguous),
1353		SUBTEST(igt_mock_splintered_region),
1354		SUBTEST(igt_mock_max_segment),
1355		SUBTEST(igt_mock_io_size),
1356	};
1357	struct intel_memory_region *mem;
1358	struct drm_i915_private *i915;
1359	int err;
1360
1361	i915 = mock_gem_device();
1362	if (!i915)
1363		return -ENOMEM;
1364
1365	mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
1366	if (IS_ERR(mem)) {
1367		pr_err("failed to create memory region\n");
1368		err = PTR_ERR(mem);
1369		goto out_unref;
1370	}
1371
1372	err = i915_subtests(tests, mem);
1373
1374	intel_memory_region_destroy(mem);
1375out_unref:
1376	mock_destroy_device(i915);
1377	return err;
1378}
1379
1380int intel_memory_region_live_selftests(struct drm_i915_private *i915)
1381{
1382	static const struct i915_subtest tests[] = {
1383		SUBTEST(igt_lmem_create),
1384		SUBTEST(igt_lmem_create_with_ps),
1385		SUBTEST(igt_lmem_create_cleared_cpu),
1386		SUBTEST(igt_lmem_write_cpu),
1387		SUBTEST(igt_lmem_write_gpu),
1388	};
1389
1390	if (!HAS_LMEM(i915)) {
1391		pr_info("device lacks LMEM support, skipping\n");
1392		return 0;
1393	}
1394
1395	if (intel_gt_is_wedged(to_gt(i915)))
1396		return 0;
1397
1398	return i915_live_subtests(tests, i915);
1399}
1400
1401int intel_memory_region_perf_selftests(struct drm_i915_private *i915)
1402{
1403	static const struct i915_subtest tests[] = {
1404		SUBTEST(perf_memcpy),
1405	};
1406
1407	if (intel_gt_is_wedged(to_gt(i915)))
1408		return 0;
1409
1410	return i915_live_subtests(tests, i915);
1411}
v5.9
  1// SPDX-License-Identifier: MIT
  2/*
  3 * Copyright © 2019 Intel Corporation
  4 */
  5
  6#include <linux/prime_numbers.h>
  7#include <linux/sort.h>
  8
 
 
  9#include "../i915_selftest.h"
 10
 11#include "mock_drm.h"
 12#include "mock_gem_device.h"
 13#include "mock_region.h"
 14
 15#include "gem/i915_gem_context.h"
 16#include "gem/i915_gem_lmem.h"
 17#include "gem/i915_gem_region.h"
 18#include "gem/i915_gem_object_blt.h"
 19#include "gem/selftests/igt_gem_utils.h"
 20#include "gem/selftests/mock_context.h"
 
 21#include "gt/intel_engine_user.h"
 22#include "gt/intel_gt.h"
 
 23#include "i915_memcpy.h"
 
 24#include "selftests/igt_flush_test.h"
 25#include "selftests/i915_random.h"
 26
 27static void close_objects(struct intel_memory_region *mem,
 28			  struct list_head *objects)
 29{
 30	struct drm_i915_private *i915 = mem->i915;
 31	struct drm_i915_gem_object *obj, *on;
 32
 33	list_for_each_entry_safe(obj, on, objects, st_link) {
 
 34		if (i915_gem_object_has_pinned_pages(obj))
 35			i915_gem_object_unpin_pages(obj);
 36		/* No polluting the memory region between tests */
 37		__i915_gem_object_put_pages(obj);
 
 38		list_del(&obj->st_link);
 39		i915_gem_object_put(obj);
 40	}
 41
 42	cond_resched();
 43
 44	i915_gem_drain_freed_objects(i915);
 45}
 46
 47static int igt_mock_fill(void *arg)
 48{
 49	struct intel_memory_region *mem = arg;
 50	resource_size_t total = resource_size(&mem->region);
 51	resource_size_t page_size;
 52	resource_size_t rem;
 53	unsigned long max_pages;
 54	unsigned long page_num;
 55	LIST_HEAD(objects);
 56	int err = 0;
 57
 58	page_size = mem->mm.chunk_size;
 59	max_pages = div64_u64(total, page_size);
 60	rem = total;
 61
 62	for_each_prime_number_from(page_num, 1, max_pages) {
 63		resource_size_t size = page_num * page_size;
 64		struct drm_i915_gem_object *obj;
 65
 66		obj = i915_gem_object_create_region(mem, size, 0);
 67		if (IS_ERR(obj)) {
 68			err = PTR_ERR(obj);
 69			break;
 70		}
 71
 72		err = i915_gem_object_pin_pages(obj);
 73		if (err) {
 74			i915_gem_object_put(obj);
 75			break;
 76		}
 77
 78		list_add(&obj->st_link, &objects);
 79		rem -= size;
 80	}
 81
 82	if (err == -ENOMEM)
 83		err = 0;
 84	if (err == -ENXIO) {
 85		if (page_num * page_size <= rem) {
 86			pr_err("%s failed, space still left in region\n",
 87			       __func__);
 88			err = -EINVAL;
 89		} else {
 90			err = 0;
 91		}
 92	}
 93
 94	close_objects(mem, &objects);
 95
 96	return err;
 97}
 98
 99static struct drm_i915_gem_object *
100igt_object_create(struct intel_memory_region *mem,
101		  struct list_head *objects,
102		  u64 size,
103		  unsigned int flags)
104{
105	struct drm_i915_gem_object *obj;
106	int err;
107
108	obj = i915_gem_object_create_region(mem, size, flags);
109	if (IS_ERR(obj))
110		return obj;
111
112	err = i915_gem_object_pin_pages(obj);
113	if (err)
114		goto put;
115
116	list_add(&obj->st_link, objects);
117	return obj;
118
119put:
120	i915_gem_object_put(obj);
121	return ERR_PTR(err);
122}
123
124static void igt_object_release(struct drm_i915_gem_object *obj)
125{
 
126	i915_gem_object_unpin_pages(obj);
127	__i915_gem_object_put_pages(obj);
 
128	list_del(&obj->st_link);
129	i915_gem_object_put(obj);
130}
131
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
132static int igt_mock_contiguous(void *arg)
133{
134	struct intel_memory_region *mem = arg;
135	struct drm_i915_gem_object *obj;
136	unsigned long n_objects;
137	LIST_HEAD(objects);
138	LIST_HEAD(holes);
139	I915_RND_STATE(prng);
140	resource_size_t total;
141	resource_size_t min;
142	u64 target;
143	int err = 0;
144
145	total = resource_size(&mem->region);
146
147	/* Min size */
148	obj = igt_object_create(mem, &objects, mem->mm.chunk_size,
149				I915_BO_ALLOC_CONTIGUOUS);
150	if (IS_ERR(obj))
151		return PTR_ERR(obj);
152
153	if (obj->mm.pages->nents != 1) {
154		pr_err("%s min object spans multiple sg entries\n", __func__);
155		err = -EINVAL;
156		goto err_close_objects;
157	}
158
159	igt_object_release(obj);
160
161	/* Max size */
162	obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
163	if (IS_ERR(obj))
164		return PTR_ERR(obj);
165
166	if (obj->mm.pages->nents != 1) {
167		pr_err("%s max object spans multiple sg entries\n", __func__);
168		err = -EINVAL;
169		goto err_close_objects;
170	}
171
172	igt_object_release(obj);
173
174	/* Internal fragmentation should not bleed into the object size */
175	target = i915_prandom_u64_state(&prng);
176	div64_u64_rem(target, total, &target);
177	target = round_up(target, PAGE_SIZE);
178	target = max_t(u64, PAGE_SIZE, target);
179
180	obj = igt_object_create(mem, &objects, target,
181				I915_BO_ALLOC_CONTIGUOUS);
182	if (IS_ERR(obj))
183		return PTR_ERR(obj);
184
185	if (obj->base.size != target) {
186		pr_err("%s obj->base.size(%zx) != target(%llx)\n", __func__,
187		       obj->base.size, target);
188		err = -EINVAL;
189		goto err_close_objects;
190	}
191
192	if (obj->mm.pages->nents != 1) {
193		pr_err("%s object spans multiple sg entries\n", __func__);
194		err = -EINVAL;
195		goto err_close_objects;
196	}
197
198	igt_object_release(obj);
199
200	/*
201	 * Try to fragment the address space, such that half of it is free, but
202	 * the max contiguous block size is SZ_64K.
203	 */
204
205	target = SZ_64K;
206	n_objects = div64_u64(total, target);
207
208	while (n_objects--) {
209		struct list_head *list;
210
211		if (n_objects % 2)
212			list = &holes;
213		else
214			list = &objects;
215
216		obj = igt_object_create(mem, list, target,
217					I915_BO_ALLOC_CONTIGUOUS);
218		if (IS_ERR(obj)) {
219			err = PTR_ERR(obj);
220			goto err_close_objects;
221		}
222	}
223
224	close_objects(mem, &holes);
225
226	min = target;
227	target = total >> 1;
228
229	/* Make sure we can still allocate all the fragmented space */
230	obj = igt_object_create(mem, &objects, target, 0);
231	if (IS_ERR(obj)) {
232		err = PTR_ERR(obj);
233		goto err_close_objects;
234	}
235
236	igt_object_release(obj);
237
238	/*
239	 * Even though we have enough free space, we don't have a big enough
240	 * contiguous block. Make sure that holds true.
241	 */
242
243	do {
244		bool should_fail = target > min;
245
246		obj = igt_object_create(mem, &objects, target,
247					I915_BO_ALLOC_CONTIGUOUS);
248		if (should_fail != IS_ERR(obj)) {
249			pr_err("%s target allocation(%llx) mismatch\n",
250			       __func__, target);
251			err = -EINVAL;
252			goto err_close_objects;
253		}
254
255		target >>= 1;
256	} while (target >= mem->mm.chunk_size);
257
258err_close_objects:
259	list_splice_tail(&holes, &objects);
260	close_objects(mem, &objects);
261	return err;
262}
263
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
264static int igt_gpu_write_dw(struct intel_context *ce,
265			    struct i915_vma *vma,
266			    u32 dword,
267			    u32 value)
268{
269	return igt_gpu_fill_dw(ce, vma, dword * sizeof(u32),
270			       vma->size >> PAGE_SHIFT, value);
271}
272
273static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
274{
275	unsigned long n = obj->base.size >> PAGE_SHIFT;
276	u32 *ptr;
277	int err;
278
279	err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
280	if (err)
281		return err;
282
283	ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
284	if (IS_ERR(ptr))
285		return PTR_ERR(ptr);
286
287	ptr += dword;
288	while (n--) {
289		if (*ptr != val) {
290			pr_err("base[%u]=%08x, val=%08x\n",
291			       dword, *ptr, val);
292			err = -EINVAL;
293			break;
294		}
295
296		ptr += PAGE_SIZE / sizeof(*ptr);
297	}
298
299	i915_gem_object_unpin_map(obj);
300	return err;
301}
302
303static int igt_gpu_write(struct i915_gem_context *ctx,
304			 struct drm_i915_gem_object *obj)
305{
306	struct i915_gem_engines *engines;
307	struct i915_gem_engines_iter it;
308	struct i915_address_space *vm;
309	struct intel_context *ce;
310	I915_RND_STATE(prng);
311	IGT_TIMEOUT(end_time);
312	unsigned int count;
313	struct i915_vma *vma;
314	int *order;
315	int i, n;
316	int err = 0;
317
318	GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
319
320	n = 0;
321	count = 0;
322	for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
323		count++;
324		if (!intel_engine_can_store_dword(ce->engine))
325			continue;
326
327		vm = ce->vm;
328		n++;
329	}
330	i915_gem_context_unlock_engines(ctx);
331	if (!n)
332		return 0;
333
334	order = i915_random_order(count * count, &prng);
335	if (!order)
336		return -ENOMEM;
337
338	vma = i915_vma_instance(obj, vm, NULL);
339	if (IS_ERR(vma)) {
340		err = PTR_ERR(vma);
341		goto out_free;
342	}
343
344	err = i915_vma_pin(vma, 0, 0, PIN_USER);
345	if (err)
346		goto out_free;
347
348	i = 0;
349	engines = i915_gem_context_lock_engines(ctx);
350	do {
351		u32 rng = prandom_u32_state(&prng);
352		u32 dword = offset_in_page(rng) / 4;
353
354		ce = engines->engines[order[i] % engines->num_engines];
355		i = (i + 1) % (count * count);
356		if (!ce || !intel_engine_can_store_dword(ce->engine))
357			continue;
358
359		err = igt_gpu_write_dw(ce, vma, dword, rng);
360		if (err)
361			break;
362
 
363		err = igt_cpu_check(obj, dword, rng);
 
364		if (err)
365			break;
366	} while (!__igt_timeout(end_time, NULL));
367	i915_gem_context_unlock_engines(ctx);
368
369out_free:
370	kfree(order);
371
372	if (err == -ENOMEM)
373		err = 0;
374
375	return err;
376}
377
378static int igt_lmem_create(void *arg)
379{
380	struct drm_i915_private *i915 = arg;
381	struct drm_i915_gem_object *obj;
382	int err = 0;
383
384	obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
385	if (IS_ERR(obj))
386		return PTR_ERR(obj);
387
388	err = i915_gem_object_pin_pages(obj);
389	if (err)
390		goto out_put;
391
392	i915_gem_object_unpin_pages(obj);
393out_put:
394	i915_gem_object_put(obj);
395
396	return err;
397}
398
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
399static int igt_lmem_write_gpu(void *arg)
400{
401	struct drm_i915_private *i915 = arg;
402	struct drm_i915_gem_object *obj;
403	struct i915_gem_context *ctx;
404	struct file *file;
405	I915_RND_STATE(prng);
406	u32 sz;
407	int err;
408
409	file = mock_file(i915);
410	if (IS_ERR(file))
411		return PTR_ERR(file);
412
413	ctx = live_context(i915, file);
414	if (IS_ERR(ctx)) {
415		err = PTR_ERR(ctx);
416		goto out_file;
417	}
418
419	sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
420
421	obj = i915_gem_object_create_lmem(i915, sz, 0);
422	if (IS_ERR(obj)) {
423		err = PTR_ERR(obj);
424		goto out_file;
425	}
426
427	err = i915_gem_object_pin_pages(obj);
428	if (err)
429		goto out_put;
430
431	err = igt_gpu_write(ctx, obj);
432	if (err)
433		pr_err("igt_gpu_write failed(%d)\n", err);
434
435	i915_gem_object_unpin_pages(obj);
436out_put:
437	i915_gem_object_put(obj);
438out_file:
439	fput(file);
440	return err;
441}
442
443static struct intel_engine_cs *
444random_engine_class(struct drm_i915_private *i915,
445		    unsigned int class,
446		    struct rnd_state *prng)
447{
448	struct intel_engine_cs *engine;
449	unsigned int count;
450
451	count = 0;
452	for (engine = intel_engine_lookup_user(i915, class, 0);
453	     engine && engine->uabi_class == class;
454	     engine = rb_entry_safe(rb_next(&engine->uabi_node),
455				    typeof(*engine), uabi_node))
456		count++;
457
458	count = i915_prandom_u32_max_state(count, prng);
459	return intel_engine_lookup_user(i915, class, count);
460}
461
462static int igt_lmem_write_cpu(void *arg)
463{
464	struct drm_i915_private *i915 = arg;
465	struct drm_i915_gem_object *obj;
466	I915_RND_STATE(prng);
467	IGT_TIMEOUT(end_time);
468	u32 bytes[] = {
469		0, /* rng placeholder */
470		sizeof(u32),
471		sizeof(u64),
472		64, /* cl */
473		PAGE_SIZE,
474		PAGE_SIZE - sizeof(u32),
475		PAGE_SIZE - sizeof(u64),
476		PAGE_SIZE - 64,
477	};
478	struct intel_engine_cs *engine;
 
479	u32 *vaddr;
480	u32 sz;
481	u32 i;
482	int *order;
483	int count;
484	int err;
485
486	engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
487	if (!engine)
488		return 0;
489
490	pr_info("%s: using %s\n", __func__, engine->name);
491
492	sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
493	sz = max_t(u32, 2 * PAGE_SIZE, sz);
494
495	obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
496	if (IS_ERR(obj))
497		return PTR_ERR(obj);
498
499	vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
500	if (IS_ERR(vaddr)) {
501		err = PTR_ERR(vaddr);
502		goto out_put;
503	}
504
 
 
 
 
 
 
 
 
505	/* Put the pages into a known state -- from the gpu for added fun */
506	intel_engine_pm_get(engine);
507	err = i915_gem_object_fill_blt(obj, engine->kernel_context, 0xdeadbeaf);
 
 
 
 
 
 
 
 
508	intel_engine_pm_put(engine);
509	if (err)
510		goto out_unpin;
511
512	i915_gem_object_lock(obj);
513	err = i915_gem_object_set_to_wc_domain(obj, true);
514	i915_gem_object_unlock(obj);
515	if (err)
516		goto out_unpin;
517
518	count = ARRAY_SIZE(bytes);
519	order = i915_random_order(count * count, &prng);
520	if (!order) {
521		err = -ENOMEM;
522		goto out_unpin;
523	}
524
525	/* We want to throw in a random width/align */
526	bytes[0] = igt_random_offset(&prng, 0, PAGE_SIZE, sizeof(u32),
527				     sizeof(u32));
528
529	i = 0;
530	do {
531		u32 offset;
532		u32 align;
533		u32 dword;
534		u32 size;
535		u32 val;
536
537		size = bytes[order[i] % count];
538		i = (i + 1) % (count * count);
539
540		align = bytes[order[i] % count];
541		i = (i + 1) % (count * count);
542
543		align = max_t(u32, sizeof(u32), rounddown_pow_of_two(align));
544
545		offset = igt_random_offset(&prng, 0, obj->base.size,
546					   size, align);
547
548		val = prandom_u32_state(&prng);
549		memset32(vaddr + offset / sizeof(u32), val ^ 0xdeadbeaf,
550			 size / sizeof(u32));
551
552		/*
553		 * Sample random dw -- don't waste precious time reading every
554		 * single dw.
555		 */
556		dword = igt_random_offset(&prng, offset,
557					  offset + size,
558					  sizeof(u32), sizeof(u32));
559		dword /= sizeof(u32);
560		if (vaddr[dword] != (val ^ 0xdeadbeaf)) {
561			pr_err("%s vaddr[%u]=%u, val=%u, size=%u, align=%u, offset=%u\n",
562			       __func__, dword, vaddr[dword], val ^ 0xdeadbeaf,
563			       size, align, offset);
564			err = -EINVAL;
565			break;
566		}
567	} while (!__igt_timeout(end_time, NULL));
568
569out_unpin:
570	i915_gem_object_unpin_map(obj);
571out_put:
572	i915_gem_object_put(obj);
573
574	return err;
575}
576
577static const char *repr_type(u32 type)
578{
579	switch (type) {
580	case I915_MAP_WB:
581		return "WB";
582	case I915_MAP_WC:
583		return "WC";
584	}
585
586	return "";
587}
588
589static struct drm_i915_gem_object *
590create_region_for_mapping(struct intel_memory_region *mr, u64 size, u32 type,
591			  void **out_addr)
592{
593	struct drm_i915_gem_object *obj;
594	void *addr;
595
596	obj = i915_gem_object_create_region(mr, size, 0);
597	if (IS_ERR(obj)) {
598		if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
599			return ERR_PTR(-ENODEV);
600		return obj;
601	}
602
603	addr = i915_gem_object_pin_map(obj, type);
604	if (IS_ERR(addr)) {
605		i915_gem_object_put(obj);
606		if (PTR_ERR(addr) == -ENXIO)
607			return ERR_PTR(-ENODEV);
608		return addr;
609	}
610
611	*out_addr = addr;
612	return obj;
613}
614
615static int wrap_ktime_compare(const void *A, const void *B)
616{
617	const ktime_t *a = A, *b = B;
618
619	return ktime_compare(*a, *b);
620}
621
622static void igt_memcpy_long(void *dst, const void *src, size_t size)
623{
624	unsigned long *tmp = dst;
625	const unsigned long *s = src;
626
627	size = size / sizeof(unsigned long);
628	while (size--)
629		*tmp++ = *s++;
630}
631
632static inline void igt_memcpy(void *dst, const void *src, size_t size)
633{
634	memcpy(dst, src, size);
635}
636
637static inline void igt_memcpy_from_wc(void *dst, const void *src, size_t size)
638{
639	i915_memcpy_from_wc(dst, src, size);
640}
641
642static int _perf_memcpy(struct intel_memory_region *src_mr,
643			struct intel_memory_region *dst_mr,
644			u64 size, u32 src_type, u32 dst_type)
645{
646	struct drm_i915_private *i915 = src_mr->i915;
647	const struct {
648		const char *name;
649		void (*copy)(void *dst, const void *src, size_t size);
650		bool skip;
651	} tests[] = {
652		{
653			"memcpy",
654			igt_memcpy,
655		},
656		{
657			"memcpy_long",
658			igt_memcpy_long,
659		},
660		{
661			"memcpy_from_wc",
662			igt_memcpy_from_wc,
663			!i915_has_memcpy_from_wc(),
664		},
665	};
666	struct drm_i915_gem_object *src, *dst;
667	void *src_addr, *dst_addr;
668	int ret = 0;
669	int i;
670
671	src = create_region_for_mapping(src_mr, size, src_type, &src_addr);
672	if (IS_ERR(src)) {
673		ret = PTR_ERR(src);
674		goto out;
675	}
676
677	dst = create_region_for_mapping(dst_mr, size, dst_type, &dst_addr);
678	if (IS_ERR(dst)) {
679		ret = PTR_ERR(dst);
680		goto out_unpin_src;
681	}
682
683	for (i = 0; i < ARRAY_SIZE(tests); ++i) {
684		ktime_t t[5];
685		int pass;
686
687		if (tests[i].skip)
688			continue;
689
690		for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
691			ktime_t t0, t1;
692
693			t0 = ktime_get();
694
695			tests[i].copy(dst_addr, src_addr, size);
696
697			t1 = ktime_get();
698			t[pass] = ktime_sub(t1, t0);
699		}
700
701		sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
 
 
 
 
 
 
 
 
 
 
 
702		pr_info("%s src(%s, %s) -> dst(%s, %s) %14s %4llu KiB copy: %5lld MiB/s\n",
703			__func__,
704			src_mr->name,
705			repr_type(src_type),
706			dst_mr->name,
707			repr_type(dst_type),
708			tests[i].name,
709			size >> 10,
710			div64_u64(mul_u32_u32(4 * size,
711					      1000 * 1000 * 1000),
712				  t[1] + 2 * t[2] + t[3]) >> 20);
713
714		cond_resched();
715	}
716
717	i915_gem_object_unpin_map(dst);
718	i915_gem_object_put(dst);
719out_unpin_src:
720	i915_gem_object_unpin_map(src);
721	i915_gem_object_put(src);
722
723	i915_gem_drain_freed_objects(i915);
724out:
725	if (ret == -ENODEV)
726		ret = 0;
727
728	return ret;
729}
730
731static int perf_memcpy(void *arg)
732{
733	struct drm_i915_private *i915 = arg;
734	static const u32 types[] = {
735		I915_MAP_WB,
736		I915_MAP_WC,
737	};
738	static const u32 sizes[] = {
739		SZ_4K,
740		SZ_64K,
741		SZ_4M,
742	};
743	struct intel_memory_region *src_mr, *dst_mr;
744	int src_id, dst_id;
745	int i, j, k;
746	int ret;
747
748	for_each_memory_region(src_mr, i915, src_id) {
749		for_each_memory_region(dst_mr, i915, dst_id) {
750			for (i = 0; i < ARRAY_SIZE(sizes); ++i) {
751				for (j = 0; j < ARRAY_SIZE(types); ++j) {
752					for (k = 0; k < ARRAY_SIZE(types); ++k) {
753						ret = _perf_memcpy(src_mr,
754								   dst_mr,
755								   sizes[i],
756								   types[j],
757								   types[k]);
758						if (ret)
759							return ret;
760					}
761				}
762			}
763		}
764	}
765
766	return 0;
767}
768
769int intel_memory_region_mock_selftests(void)
770{
771	static const struct i915_subtest tests[] = {
 
772		SUBTEST(igt_mock_fill),
773		SUBTEST(igt_mock_contiguous),
 
 
 
774	};
775	struct intel_memory_region *mem;
776	struct drm_i915_private *i915;
777	int err;
778
779	i915 = mock_gem_device();
780	if (!i915)
781		return -ENOMEM;
782
783	mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0);
784	if (IS_ERR(mem)) {
785		pr_err("failed to create memory region\n");
786		err = PTR_ERR(mem);
787		goto out_unref;
788	}
789
790	err = i915_subtests(tests, mem);
791
792	intel_memory_region_put(mem);
793out_unref:
794	drm_dev_put(&i915->drm);
795	return err;
796}
797
798int intel_memory_region_live_selftests(struct drm_i915_private *i915)
799{
800	static const struct i915_subtest tests[] = {
801		SUBTEST(igt_lmem_create),
 
 
802		SUBTEST(igt_lmem_write_cpu),
803		SUBTEST(igt_lmem_write_gpu),
804	};
805
806	if (!HAS_LMEM(i915)) {
807		pr_info("device lacks LMEM support, skipping\n");
808		return 0;
809	}
810
811	if (intel_gt_is_wedged(&i915->gt))
812		return 0;
813
814	return i915_live_subtests(tests, i915);
815}
816
817int intel_memory_region_perf_selftests(struct drm_i915_private *i915)
818{
819	static const struct i915_subtest tests[] = {
820		SUBTEST(perf_memcpy),
821	};
822
823	if (intel_gt_is_wedged(&i915->gt))
824		return 0;
825
826	return i915_live_subtests(tests, i915);
827}