Linux Audio

Check our new training course

Loading...
Note: File does not exist in v4.6.
  1// SPDX-License-Identifier: MIT
  2/*
  3 * Copyright © 2014-2018 Intel Corporation
  4 */
  5
  6#include "gem/i915_gem_object.h"
  7
  8#include "i915_drv.h"
  9#include "intel_engine_pm.h"
 10#include "intel_gt_buffer_pool.h"
 11
 12static struct intel_gt *to_gt(struct intel_gt_buffer_pool *pool)
 13{
 14	return container_of(pool, struct intel_gt, buffer_pool);
 15}
 16
 17static struct list_head *
 18bucket_for_size(struct intel_gt_buffer_pool *pool, size_t sz)
 19{
 20	int n;
 21
 22	/*
 23	 * Compute a power-of-two bucket, but throw everything greater than
 24	 * 16KiB into the same bucket: i.e. the buckets hold objects of
 25	 * (1 page, 2 pages, 4 pages, 8+ pages).
 26	 */
 27	n = fls(sz >> PAGE_SHIFT) - 1;
 28	if (n >= ARRAY_SIZE(pool->cache_list))
 29		n = ARRAY_SIZE(pool->cache_list) - 1;
 30
 31	return &pool->cache_list[n];
 32}
 33
 34static void node_free(struct intel_gt_buffer_pool_node *node)
 35{
 36	i915_gem_object_put(node->obj);
 37	i915_active_fini(&node->active);
 38	kfree(node);
 39}
 40
 41static void pool_free_work(struct work_struct *wrk)
 42{
 43	struct intel_gt_buffer_pool *pool =
 44		container_of(wrk, typeof(*pool), work.work);
 45	struct intel_gt_buffer_pool_node *node, *next;
 46	unsigned long old = jiffies - HZ;
 47	bool active = false;
 48	LIST_HEAD(stale);
 49	int n;
 50
 51	/* Free buffers that have not been used in the past second */
 52	spin_lock_irq(&pool->lock);
 53	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) {
 54		struct list_head *list = &pool->cache_list[n];
 55
 56		/* Most recent at head; oldest at tail */
 57		list_for_each_entry_safe_reverse(node, next, list, link) {
 58			if (time_before(node->age, old))
 59				break;
 60
 61			list_move(&node->link, &stale);
 62		}
 63		active |= !list_empty(list);
 64	}
 65	spin_unlock_irq(&pool->lock);
 66
 67	list_for_each_entry_safe(node, next, &stale, link)
 68		node_free(node);
 69
 70	if (active)
 71		schedule_delayed_work(&pool->work,
 72				      round_jiffies_up_relative(HZ));
 73}
 74
 75static int pool_active(struct i915_active *ref)
 76{
 77	struct intel_gt_buffer_pool_node *node =
 78		container_of(ref, typeof(*node), active);
 79	struct dma_resv *resv = node->obj->base.resv;
 80	int err;
 81
 82	if (dma_resv_trylock(resv)) {
 83		dma_resv_add_excl_fence(resv, NULL);
 84		dma_resv_unlock(resv);
 85	}
 86
 87	err = i915_gem_object_pin_pages(node->obj);
 88	if (err)
 89		return err;
 90
 91	/* Hide this pinned object from the shrinker until retired */
 92	i915_gem_object_make_unshrinkable(node->obj);
 93
 94	return 0;
 95}
 96
 97__i915_active_call
 98static void pool_retire(struct i915_active *ref)
 99{
100	struct intel_gt_buffer_pool_node *node =
101		container_of(ref, typeof(*node), active);
102	struct intel_gt_buffer_pool *pool = node->pool;
103	struct list_head *list = bucket_for_size(pool, node->obj->base.size);
104	unsigned long flags;
105
106	i915_gem_object_unpin_pages(node->obj);
107
108	/* Return this object to the shrinker pool */
109	i915_gem_object_make_purgeable(node->obj);
110
111	spin_lock_irqsave(&pool->lock, flags);
112	node->age = jiffies;
113	list_add(&node->link, list);
114	spin_unlock_irqrestore(&pool->lock, flags);
115
116	schedule_delayed_work(&pool->work,
117			      round_jiffies_up_relative(HZ));
118}
119
120static struct intel_gt_buffer_pool_node *
121node_create(struct intel_gt_buffer_pool *pool, size_t sz)
122{
123	struct intel_gt *gt = to_gt(pool);
124	struct intel_gt_buffer_pool_node *node;
125	struct drm_i915_gem_object *obj;
126
127	node = kmalloc(sizeof(*node),
128		       GFP_KERNEL | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
129	if (!node)
130		return ERR_PTR(-ENOMEM);
131
132	node->pool = pool;
133	i915_active_init(&node->active, pool_active, pool_retire);
134
135	obj = i915_gem_object_create_internal(gt->i915, sz);
136	if (IS_ERR(obj)) {
137		i915_active_fini(&node->active);
138		kfree(node);
139		return ERR_CAST(obj);
140	}
141
142	i915_gem_object_set_readonly(obj);
143
144	node->obj = obj;
145	return node;
146}
147
148struct intel_gt_buffer_pool_node *
149intel_gt_get_buffer_pool(struct intel_gt *gt, size_t size)
150{
151	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
152	struct intel_gt_buffer_pool_node *node;
153	struct list_head *list;
154	unsigned long flags;
155	int ret;
156
157	size = PAGE_ALIGN(size);
158	list = bucket_for_size(pool, size);
159
160	spin_lock_irqsave(&pool->lock, flags);
161	list_for_each_entry(node, list, link) {
162		if (node->obj->base.size < size)
163			continue;
164		list_del(&node->link);
165		break;
166	}
167	spin_unlock_irqrestore(&pool->lock, flags);
168
169	if (&node->link == list) {
170		node = node_create(pool, size);
171		if (IS_ERR(node))
172			return node;
173	}
174
175	ret = i915_active_acquire(&node->active);
176	if (ret) {
177		node_free(node);
178		return ERR_PTR(ret);
179	}
180
181	return node;
182}
183
184void intel_gt_init_buffer_pool(struct intel_gt *gt)
185{
186	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
187	int n;
188
189	spin_lock_init(&pool->lock);
190	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
191		INIT_LIST_HEAD(&pool->cache_list[n]);
192	INIT_DELAYED_WORK(&pool->work, pool_free_work);
193}
194
195static void pool_free_imm(struct intel_gt_buffer_pool *pool)
196{
197	int n;
198
199	spin_lock_irq(&pool->lock);
200	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) {
201		struct intel_gt_buffer_pool_node *node, *next;
202		struct list_head *list = &pool->cache_list[n];
203
204		list_for_each_entry_safe(node, next, list, link)
205			node_free(node);
206		INIT_LIST_HEAD(list);
207	}
208	spin_unlock_irq(&pool->lock);
209}
210
211void intel_gt_flush_buffer_pool(struct intel_gt *gt)
212{
213	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
214
215	do {
216		pool_free_imm(pool);
217	} while (cancel_delayed_work_sync(&pool->work));
218}
219
220void intel_gt_fini_buffer_pool(struct intel_gt *gt)
221{
222	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
223	int n;
224
225	intel_gt_flush_buffer_pool(gt);
226
227	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
228		GEM_BUG_ON(!list_empty(&pool->cache_list[n]));
229}