lima_gem.c 9.09 KB
Newer Older
1 2 3
// SPDX-License-Identifier: GPL-2.0 OR MIT
/* Copyright 2017-2019 Qiang Yu <yuq825@gmail.com> */

4
#include <linux/mm.h>
5
#include <linux/iosys-map.h>
6
#include <linux/sync_file.h>
7
#include <linux/pagemap.h>
8 9
#include <linux/shmem_fs.h>
#include <linux/dma-mapping.h>
10 11 12 13 14 15 16 17 18 19 20

#include <drm/drm_file.h>
#include <drm/drm_syncobj.h>
#include <drm/drm_utils.h>

#include <drm/lima_drm.h>

#include "lima_drv.h"
#include "lima_gem.h"
#include "lima_vm.h"

21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36
int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm)
{
	struct page **pages;
	struct address_space *mapping = bo->base.base.filp->f_mapping;
	struct device *dev = bo->base.base.dev->dev;
	size_t old_size = bo->heap_size;
	size_t new_size = bo->heap_size ? bo->heap_size * 2 :
		(lima_heap_init_nr_pages << PAGE_SHIFT);
	struct sg_table sgt;
	int i, ret;

	if (bo->heap_size >= bo->base.base.size)
		return -ENOSPC;

	new_size = min(new_size, bo->base.base.size);

37
	dma_resv_lock(bo->base.base.resv, NULL);
38 39 40 41 42 43 44

	if (bo->base.pages) {
		pages = bo->base.pages;
	} else {
		pages = kvmalloc_array(bo->base.base.size >> PAGE_SHIFT,
				       sizeof(*pages), GFP_KERNEL | __GFP_ZERO);
		if (!pages) {
45
			dma_resv_unlock(bo->base.base.resv);
46 47 48 49 50 51 52 53 54 55 56 57 58
			return -ENOMEM;
		}

		bo->base.pages = pages;
		bo->base.pages_use_count = 1;

		mapping_set_unevictable(mapping);
	}

	for (i = old_size >> PAGE_SHIFT; i < new_size >> PAGE_SHIFT; i++) {
		struct page *page = shmem_read_mapping_page(mapping, i);

		if (IS_ERR(page)) {
59
			dma_resv_unlock(bo->base.base.resv);
60 61 62 63 64
			return PTR_ERR(page);
		}
		pages[i] = page;
	}

65
	dma_resv_unlock(bo->base.base.resv);
66 67 68 69 70 71 72

	ret = sg_alloc_table_from_pages(&sgt, pages, i, 0,
					new_size, GFP_KERNEL);
	if (ret)
		return ret;

	if (bo->base.sgt) {
73
		dma_unmap_sgtable(dev, bo->base.sgt, DMA_BIDIRECTIONAL, 0);
74 75 76 77
		sg_free_table(bo->base.sgt);
	} else {
		bo->base.sgt = kmalloc(sizeof(*bo->base.sgt), GFP_KERNEL);
		if (!bo->base.sgt) {
78 79
			ret = -ENOMEM;
			goto err_out0;
80 81 82
		}
	}

83
	ret = dma_map_sgtable(dev, &sgt, DMA_BIDIRECTIONAL, 0);
84 85
	if (ret)
		goto err_out1;
86 87 88 89 90 91

	*bo->base.sgt = sgt;

	if (vm) {
		ret = lima_vm_map_bo(vm, bo, old_size >> PAGE_SHIFT);
		if (ret)
92
			goto err_out2;
93 94 95 96
	}

	bo->heap_size = new_size;
	return 0;
97 98 99 100 101 102 103 104 105

err_out2:
	dma_unmap_sgtable(dev, &sgt, DMA_BIDIRECTIONAL, 0);
err_out1:
	kfree(bo->base.sgt);
	bo->base.sgt = NULL;
err_out0:
	sg_free_table(&sgt);
	return ret;
106 107
}

108 109 110 111
int lima_gem_create_handle(struct drm_device *dev, struct drm_file *file,
			   u32 size, u32 flags, u32 *handle)
{
	int err;
112 113 114
	gfp_t mask;
	struct drm_gem_shmem_object *shmem;
	struct drm_gem_object *obj;
115 116
	struct lima_bo *bo;
	bool is_heap = flags & LIMA_BO_FLAG_HEAP;
117 118 119 120

	shmem = drm_gem_shmem_create(dev, size);
	if (IS_ERR(shmem))
		return PTR_ERR(shmem);
121

122
	obj = &shmem->base;
123

124 125 126 127 128 129
	/* Mali Utgard GPU can only support 32bit address space */
	mask = mapping_gfp_mask(obj->filp->f_mapping);
	mask &= ~__GFP_HIGHMEM;
	mask |= __GFP_DMA32;
	mapping_set_gfp_mask(obj->filp->f_mapping, mask);

130 131 132 133 134 135
	if (is_heap) {
		bo = to_lima_bo(obj);
		err = lima_heap_alloc(bo, NULL);
		if (err)
			goto out;
	} else {
136
		struct sg_table *sgt = drm_gem_shmem_get_pages_sgt(shmem);
137 138 139 140 141

		if (IS_ERR(sgt)) {
			err = PTR_ERR(sgt);
			goto out;
		}
142
	}
143

144 145 146
	err = drm_gem_handle_create(file, obj, handle);

out:
147
	/* drop reference from allocate - handle holds it now */
148
	drm_gem_object_put(obj);
149 150 151 152

	return err;
}

153
static void lima_gem_free_object(struct drm_gem_object *obj)
154 155 156 157 158 159
{
	struct lima_bo *bo = to_lima_bo(obj);

	if (!list_empty(&bo->va))
		dev_err(obj->dev->dev, "lima gem free bo still has va\n");

160
	drm_gem_shmem_free(&bo->base);
161 162
}

163
static int lima_gem_object_open(struct drm_gem_object *obj, struct drm_file *file)
164 165 166 167 168 169 170 171
{
	struct lima_bo *bo = to_lima_bo(obj);
	struct lima_drm_priv *priv = to_lima_drm_priv(file);
	struct lima_vm *vm = priv->vm;

	return lima_vm_bo_add(vm, bo, true);
}

172
static void lima_gem_object_close(struct drm_gem_object *obj, struct drm_file *file)
173 174 175 176 177 178 179 180
{
	struct lima_bo *bo = to_lima_bo(obj);
	struct lima_drm_priv *priv = to_lima_drm_priv(file);
	struct lima_vm *vm = priv->vm;

	lima_vm_bo_del(vm, bo);
}

181 182 183 184 185 186 187
static int lima_gem_pin(struct drm_gem_object *obj)
{
	struct lima_bo *bo = to_lima_bo(obj);

	if (bo->heap_size)
		return -EINVAL;

188
	return drm_gem_shmem_pin_locked(&bo->base);
189 190
}

191
static int lima_gem_vmap(struct drm_gem_object *obj, struct iosys_map *map)
192 193 194 195
{
	struct lima_bo *bo = to_lima_bo(obj);

	if (bo->heap_size)
196
		return -EINVAL;
197

198
	return drm_gem_shmem_vmap(&bo->base, map);
199 200 201 202 203 204 205 206 207
}

static int lima_gem_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma)
{
	struct lima_bo *bo = to_lima_bo(obj);

	if (bo->heap_size)
		return -EINVAL;

208
	return drm_gem_shmem_mmap(&bo->base, vma);
209 210
}

211 212 213 214
static const struct drm_gem_object_funcs lima_gem_funcs = {
	.free = lima_gem_free_object,
	.open = lima_gem_object_open,
	.close = lima_gem_object_close,
215
	.print_info = drm_gem_shmem_object_print_info,
216
	.pin = lima_gem_pin,
217 218
	.unpin = drm_gem_shmem_object_unpin,
	.get_sg_table = drm_gem_shmem_object_get_sg_table,
219
	.vmap = lima_gem_vmap,
220
	.vunmap = drm_gem_shmem_object_vunmap,
221
	.mmap = lima_gem_mmap,
222
	.vm_ops = &drm_gem_shmem_vm_ops,
223 224 225 226 227 228 229 230
};

struct drm_gem_object *lima_gem_create_object(struct drm_device *dev, size_t size)
{
	struct lima_bo *bo;

	bo = kzalloc(sizeof(*bo), GFP_KERNEL);
	if (!bo)
231
		return ERR_PTR(-ENOMEM);
232 233 234

	mutex_init(&bo->lock);
	INIT_LIST_HEAD(&bo->va);
235
	bo->base.map_wc = true;
236 237 238 239 240
	bo->base.base.funcs = &lima_gem_funcs;

	return &bo->base.base;
}

241 242 243 244 245 246 247 248 249 250 251 252 253 254 255
int lima_gem_get_info(struct drm_file *file, u32 handle, u32 *va, u64 *offset)
{
	struct drm_gem_object *obj;
	struct lima_bo *bo;
	struct lima_drm_priv *priv = to_lima_drm_priv(file);
	struct lima_vm *vm = priv->vm;

	obj = drm_gem_object_lookup(file, handle);
	if (!obj)
		return -ENOENT;

	bo = to_lima_bo(obj);

	*va = lima_vm_get_va(vm, bo);

256
	*offset = drm_vma_node_offset_addr(&obj->vma_node);
257

258
	drm_gem_object_put(obj);
259 260 261 262 263 264
	return 0;
}

static int lima_gem_sync_bo(struct lima_sched_task *task, struct lima_bo *bo,
			    bool write, bool explicit)
{
265
	int err;
266

267 268 269
	err = dma_resv_reserve_fences(lima_bo_resv(bo), 1);
	if (err)
		return err;
270 271 272 273 274

	/* explicit sync use user passed dep fence */
	if (explicit)
		return 0;

275 276 277
	return drm_sched_job_add_implicit_dependencies(&task->base,
						       &bo->base.base,
						       write);
278 279 280 281 282 283 284 285 286 287
}

static int lima_gem_add_deps(struct drm_file *file, struct lima_submit *submit)
{
	int i, err;

	for (i = 0; i < ARRAY_SIZE(submit->in_sync); i++) {
		if (!submit->in_sync[i])
			continue;

288 289
		err = drm_sched_job_add_syncobj_dependency(&submit->task->base, file,
							   submit->in_sync[i], 0);
290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329
		if (err)
			return err;
	}

	return 0;
}

int lima_gem_submit(struct drm_file *file, struct lima_submit *submit)
{
	int i, err = 0;
	struct ww_acquire_ctx ctx;
	struct lima_drm_priv *priv = to_lima_drm_priv(file);
	struct lima_vm *vm = priv->vm;
	struct drm_syncobj *out_sync = NULL;
	struct dma_fence *fence;
	struct lima_bo **bos = submit->lbos;

	if (submit->out_sync) {
		out_sync = drm_syncobj_find(file, submit->out_sync);
		if (!out_sync)
			return -ENOENT;
	}

	for (i = 0; i < submit->nr_bos; i++) {
		struct drm_gem_object *obj;
		struct lima_bo *bo;

		obj = drm_gem_object_lookup(file, submit->bos[i].handle);
		if (!obj) {
			err = -ENOENT;
			goto err_out0;
		}

		bo = to_lima_bo(obj);

		/* increase refcnt of gpu va map to prevent unmapped when executing,
		 * will be decreased when task done
		 */
		err = lima_vm_bo_add(vm, bo, false);
		if (err) {
330
			drm_gem_object_put(obj);
331 332 333 334 335 336
			goto err_out0;
		}

		bos[i] = bo;
	}

337 338
	err = drm_gem_lock_reservations((struct drm_gem_object **)bos,
					submit->nr_bos, &ctx);
339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360
	if (err)
		goto err_out0;

	err = lima_sched_task_init(
		submit->task, submit->ctx->context + submit->pipe,
		bos, submit->nr_bos, vm);
	if (err)
		goto err_out1;

	err = lima_gem_add_deps(file, submit);
	if (err)
		goto err_out2;

	for (i = 0; i < submit->nr_bos; i++) {
		err = lima_gem_sync_bo(
			submit->task, bos[i],
			submit->bos[i].flags & LIMA_SUBMIT_BO_WRITE,
			submit->flags & LIMA_SUBMIT_FLAG_EXPLICIT_FENCE);
		if (err)
			goto err_out2;
	}

361
	fence = lima_sched_context_queue_task(submit->task);
362 363

	for (i = 0; i < submit->nr_bos; i++) {
364 365 366
		dma_resv_add_fence(lima_bo_resv(bos[i]), fence,
				   submit->bos[i].flags & LIMA_SUBMIT_BO_WRITE ?
				   DMA_RESV_USAGE_WRITE : DMA_RESV_USAGE_READ);
367 368
	}

369 370
	drm_gem_unlock_reservations((struct drm_gem_object **)bos,
				    submit->nr_bos, &ctx);
371 372

	for (i = 0; i < submit->nr_bos; i++)
373
		drm_gem_object_put(&bos[i]->base.base);
374 375 376 377 378 379 380 381 382 383 384 385 386

	if (out_sync) {
		drm_syncobj_replace_fence(out_sync, fence);
		drm_syncobj_put(out_sync);
	}

	dma_fence_put(fence);

	return 0;

err_out2:
	lima_sched_task_fini(submit->task);
err_out1:
387 388
	drm_gem_unlock_reservations((struct drm_gem_object **)bos,
				    submit->nr_bos, &ctx);
389 390 391 392 393
err_out0:
	for (i = 0; i < submit->nr_bos; i++) {
		if (!bos[i])
			break;
		lima_vm_bo_del(vm, bos[i]);
394
		drm_gem_object_put(&bos[i]->base.base);
395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410
	}
	if (out_sync)
		drm_syncobj_put(out_sync);
	return err;
}

int lima_gem_wait(struct drm_file *file, u32 handle, u32 op, s64 timeout_ns)
{
	bool write = op & LIMA_GEM_WAIT_WRITE;
	long ret, timeout;

	if (!op)
		return 0;

	timeout = drm_timeout_abs_to_jiffies(timeout_ns);

411
	ret = drm_gem_dma_resv_wait(file, handle, write, timeout);
412
	if (ret == -ETIME)
413 414 415 416
		ret = timeout ? -ETIMEDOUT : -EBUSY;

	return ret;
}