// SPDX-License-Identifier: GPL-2.0 OR MIT /* * Copyright 2011 Red Hat Inc. * Copyright 2023 Intel Corporation. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE * USE OR OTHER DEALINGS IN THE SOFTWARE. * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * */ /* Algorithm: * * We store the last allocated bo in "hole", we always try to allocate * after the last allocated bo. Principle is that in a linear GPU ring * progression was is after last is the oldest bo we allocated and thus * the first one that should no longer be in use by the GPU. * * If it's not the case we skip over the bo after last to the closest * done bo if such one exist. If none exist and we are not asked to * block we report failure to allocate. * * If we are asked to block we wait on all the oldest fence of all * rings. We just wait for any of those fence to complete. */ #include #include #include #include #include #include static void drm_suballoc_remove_locked(struct drm_suballoc *sa); static void drm_suballoc_try_free(struct drm_suballoc_manager *sa_manager); /** * drm_suballoc_manager_init() - Initialise the drm_suballoc_manager * @sa_manager: pointer to the sa_manager * @size: number of bytes we want to suballocate * @align: alignment for each suballocated chunk * * Prepares the suballocation manager for suballocations. */ void drm_suballoc_manager_init(struct drm_suballoc_manager *sa_manager, size_t size, size_t align) { unsigned int i; BUILD_BUG_ON(!is_power_of_2(DRM_SUBALLOC_MAX_QUEUES)); if (!align) align = 1; /* alignment must be a power of 2 */ if (WARN_ON_ONCE(align & (align - 1))) align = roundup_pow_of_two(align); init_waitqueue_head(&sa_manager->wq); sa_manager->size = size; sa_manager->align = align; sa_manager->hole = &sa_manager->olist; INIT_LIST_HEAD(&sa_manager->olist); for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) INIT_LIST_HEAD(&sa_manager->flist[i]); } EXPORT_SYMBOL(drm_suballoc_manager_init); /** * drm_suballoc_manager_fini() - Destroy the drm_suballoc_manager * @sa_manager: pointer to the sa_manager * * Cleans up the suballocation manager after use. All fences added * with drm_suballoc_free() must be signaled, or we cannot clean up * the entire manager. */ void drm_suballoc_manager_fini(struct drm_suballoc_manager *sa_manager) { struct drm_suballoc *sa, *tmp; if (!sa_manager->size) return; if (!list_empty(&sa_manager->olist)) { sa_manager->hole = &sa_manager->olist; drm_suballoc_try_free(sa_manager); if (!list_empty(&sa_manager->olist)) DRM_ERROR("sa_manager is not empty, clearing anyway\n"); } list_for_each_entry_safe(sa, tmp, &sa_manager->olist, olist) { drm_suballoc_remove_locked(sa); } sa_manager->size = 0; } EXPORT_SYMBOL(drm_suballoc_manager_fini); static void drm_suballoc_remove_locked(struct drm_suballoc *sa) { struct drm_suballoc_manager *sa_manager = sa->manager; if (sa_manager->hole == &sa->olist) sa_manager->hole = sa->olist.prev; list_del_init(&sa->olist); list_del_init(&sa->flist); dma_fence_put(sa->fence); kfree(sa); } static void drm_suballoc_try_free(struct drm_suballoc_manager *sa_manager) { struct drm_suballoc *sa, *tmp; if (sa_manager->hole->next == &sa_manager->olist) return; sa = list_entry(sa_manager->hole->next, struct drm_suballoc, olist); list_for_each_entry_safe_from(sa, tmp, &sa_manager->olist, olist) { if (!sa->fence || !dma_fence_is_signaled(sa->fence)) return; drm_suballoc_remove_locked(sa); } } static size_t drm_suballoc_hole_soffset(struct drm_suballoc_manager *sa_manager) { struct list_head *hole = sa_manager->hole; if (hole != &sa_manager->olist) return list_entry(hole, struct drm_suballoc, olist)->eoffset; return 0; } static size_t drm_suballoc_hole_eoffset(struct drm_suballoc_manager *sa_manager) { struct list_head *hole = sa_manager->hole; if (hole->next != &sa_manager->olist) return list_entry(hole->next, struct drm_suballoc, olist)->soffset; return sa_manager->size; } static bool drm_suballoc_try_alloc(struct drm_suballoc_manager *sa_manager, struct drm_suballoc *sa, size_t size, size_t align) { size_t soffset, eoffset, wasted; soffset = drm_suballoc_hole_soffset(sa_manager); eoffset = drm_suballoc_hole_eoffset(sa_manager); wasted = round_up(soffset, align) - soffset; if ((eoffset - soffset) >= (size + wasted)) { soffset += wasted; sa->manager = sa_manager; sa->soffset = soffset; sa->eoffset = soffset + size; list_add(&sa->olist, sa_manager->hole); INIT_LIST_HEAD(&sa->flist); sa_manager->hole = &sa->olist; return true; } return false; } static bool __drm_suballoc_event(struct drm_suballoc_manager *sa_manager, size_t size, size_t align) { size_t soffset, eoffset, wasted; unsigned int i; for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) if (!list_empty(&sa_manager->flist[i])) return true; soffset = drm_suballoc_hole_soffset(sa_manager); eoffset = drm_suballoc_hole_eoffset(sa_manager); wasted = round_up(soffset, align) - soffset; return ((eoffset - soffset) >= (size + wasted)); } /** * drm_suballoc_event() - Check if we can stop waiting * @sa_manager: pointer to the sa_manager * @size: number of bytes we want to allocate * @align: alignment we need to match * * Return: true if either there is a fence we can wait for or * enough free memory to satisfy the allocation directly. * false otherwise. */ static bool drm_suballoc_event(struct drm_suballoc_manager *sa_manager, size_t size, size_t align) { bool ret; spin_lock(&sa_manager->wq.lock); ret = __drm_suballoc_event(sa_manager, size, align); spin_unlock(&sa_manager->wq.lock); return ret; } static bool drm_suballoc_next_hole(struct drm_suballoc_manager *sa_manager, struct dma_fence **fences, unsigned int *tries) { struct drm_suballoc *best_bo = NULL; unsigned int i, best_idx; size_t soffset, best, tmp; /* if hole points to the end of the buffer */ if (sa_manager->hole->next == &sa_manager->olist) { /* try again with its beginning */ sa_manager->hole = &sa_manager->olist; return true; } soffset = drm_suballoc_hole_soffset(sa_manager); /* to handle wrap around we add sa_manager->size */ best = sa_manager->size * 2; /* go over all fence list and try to find the closest sa * of the current last */ for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) { struct drm_suballoc *sa; fences[i] = NULL; if (list_empty(&sa_manager->flist[i])) continue; sa = list_first_entry(&sa_manager->flist[i], struct drm_suballoc, flist); if (!dma_fence_is_signaled(sa->fence)) { fences[i] = sa->fence; continue; } /* limit the number of tries each freelist gets */ if (tries[i] > 2) continue; tmp = sa->soffset; if (tmp < soffset) { /* wrap around, pretend it's after */ tmp += sa_manager->size; } tmp -= soffset; if (tmp < best) { /* this sa bo is the closest one */ best = tmp; best_idx = i; best_bo = sa; } } if (best_bo) { ++tries[best_idx]; sa_manager->hole = best_bo->olist.prev; /* * We know that this one is signaled, * so it's safe to remove it. */ drm_suballoc_remove_locked(best_bo); return true; } return false; } /** * drm_suballoc_new() - Make a suballocation. * @sa_manager: pointer to the sa_manager * @size: number of bytes we want to suballocate. * @gfp: gfp flags used for memory allocation. Typically GFP_KERNEL but * the argument is provided for suballocations from reclaim context or * where the caller wants to avoid pipelining rather than wait for * reclaim. * @intr: Whether to perform waits interruptible. This should typically * always be true, unless the caller needs to propagate a * non-interruptible context from above layers. * @align: Alignment. Must not exceed the default manager alignment. * If @align is zero, then the manager alignment is used. * * Try to make a suballocation of size @size, which will be rounded * up to the alignment specified in specified in drm_suballoc_manager_init(). * * Return: a new suballocated bo, or an ERR_PTR. */ struct drm_suballoc * drm_suballoc_new(struct drm_suballoc_manager *sa_manager, size_t size, gfp_t gfp, bool intr, size_t align) { struct dma_fence *fences[DRM_SUBALLOC_MAX_QUEUES]; unsigned int tries[DRM_SUBALLOC_MAX_QUEUES]; unsigned int count; int i, r; struct drm_suballoc *sa; if (WARN_ON_ONCE(align > sa_manager->align)) return ERR_PTR(-EINVAL); if (WARN_ON_ONCE(size > sa_manager->size || !size)) return ERR_PTR(-EINVAL); if (!align) align = sa_manager->align; sa = kmalloc(sizeof(*sa), gfp); if (!sa) return ERR_PTR(-ENOMEM); sa->manager = sa_manager; sa->fence = NULL; INIT_LIST_HEAD(&sa->olist); INIT_LIST_HEAD(&sa->flist); spin_lock(&sa_manager->wq.lock); do { for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) tries[i] = 0; do { drm_suballoc_try_free(sa_manager); if (drm_suballoc_try_alloc(sa_manager, sa, size, align)) { spin_unlock(&sa_manager->wq.lock); return sa; } /* see if we can skip over some allocations */ } while (drm_suballoc_next_hole(sa_manager, fences, tries)); for (i = 0, count = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) if (fences[i]) fences[count++] = dma_fence_get(fences[i]); if (count) { long t; spin_unlock(&sa_manager->wq.lock); t = dma_fence_wait_any_timeout(fences, count, intr, MAX_SCHEDULE_TIMEOUT, NULL); for (i = 0; i < count; ++i) dma_fence_put(fences[i]); r = (t > 0) ? 0 : t; spin_lock(&sa_manager->wq.lock); } else if (intr) { /* if we have nothing to wait for block */ r = wait_event_interruptible_locked (sa_manager->wq, __drm_suballoc_event(sa_manager, size, align)); } else { spin_unlock(&sa_manager->wq.lock); wait_event(sa_manager->wq, drm_suballoc_event(sa_manager, size, align)); r = 0; spin_lock(&sa_manager->wq.lock); } } while (!r); spin_unlock(&sa_manager->wq.lock); kfree(sa); return ERR_PTR(r); } EXPORT_SYMBOL(drm_suballoc_new); /** * drm_suballoc_free - Free a suballocation * @suballoc: pointer to the suballocation * @fence: fence that signals when suballocation is idle * * Free the suballocation. The suballocation can be re-used after @fence signals. */ void drm_suballoc_free(struct drm_suballoc *suballoc, struct dma_fence *fence) { struct drm_suballoc_manager *sa_manager; if (!suballoc) return; sa_manager = suballoc->manager; spin_lock(&sa_manager->wq.lock); if (fence && !dma_fence_is_signaled(fence)) { u32 idx; suballoc->fence = dma_fence_get(fence); idx = fence->context & (DRM_SUBALLOC_MAX_QUEUES - 1); list_add_tail(&suballoc->flist, &sa_manager->flist[idx]); } else { drm_suballoc_remove_locked(suballoc); } wake_up_all_locked(&sa_manager->wq); spin_unlock(&sa_manager->wq.lock); } EXPORT_SYMBOL(drm_suballoc_free); #ifdef CONFIG_DEBUG_FS void drm_suballoc_dump_debug_info(struct drm_suballoc_manager *sa_manager, struct drm_printer *p, unsigned long long suballoc_base) { struct drm_suballoc *i; spin_lock(&sa_manager->wq.lock); list_for_each_entry(i, &sa_manager->olist, olist) { unsigned long long soffset = i->soffset; unsigned long long eoffset = i->eoffset; if (&i->olist == sa_manager->hole) drm_puts(p, ">"); else drm_puts(p, " "); drm_printf(p, "[0x%010llx 0x%010llx] size %8lld", suballoc_base + soffset, suballoc_base + eoffset, eoffset - soffset); if (i->fence) drm_printf(p, " protected by 0x%016llx on context %llu", (unsigned long long)i->fence->seqno, (unsigned long long)i->fence->context); drm_puts(p, "\n"); } spin_unlock(&sa_manager->wq.lock); } EXPORT_SYMBOL(drm_suballoc_dump_debug_info); #endif MODULE_AUTHOR("Multiple"); MODULE_DESCRIPTION("Range suballocator helper"); MODULE_LICENSE("Dual MIT/GPL");