/* SPDX-License-Identifier: GPL-2.0-only OR MIT */ /* * Copyright 2020 Advanced Micro Devices, Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. */ #ifndef _XE_RES_CURSOR_H_ #define _XE_RES_CURSOR_H_ #include #include #include #include #include #include #include "xe_bo.h" #include "xe_device.h" #include "xe_macros.h" #include "xe_svm.h" #include "xe_ttm_vram_mgr.h" /** * struct xe_res_cursor - state for walking over dma mapping, vram_mgr, * stolen_mgr, and gtt_mgr allocations */ struct xe_res_cursor { /** @start: Start of cursor */ u64 start; /** @size: Size of the current segment. */ u64 size; /** @remaining: Remaining bytes in cursor */ u64 remaining; /** @node: Opaque point current node cursor */ void *node; /** @mem_type: Memory type */ u32 mem_type; /** @sgl: Scatterlist for cursor */ struct scatterlist *sgl; /** @dma_addr: Current element in a struct drm_pagemap_device_addr array */ const struct drm_pagemap_device_addr *dma_addr; /** @mm: Buddy allocator for VRAM cursor */ struct drm_buddy *mm; /** * @dma_start: DMA start address for the current segment. * This may be different to @dma_addr.addr since elements in * the array may be coalesced to a single segment. */ u64 dma_start; /** @dma_seg_size: Size of the current DMA segment. */ u64 dma_seg_size; }; static struct drm_buddy *xe_res_get_buddy(struct ttm_resource *res) { struct ttm_resource_manager *mgr; mgr = ttm_manager_type(res->bo->bdev, res->mem_type); return &to_xe_ttm_vram_mgr(mgr)->mm; } /** * xe_res_first - initialize a xe_res_cursor * * @res: TTM resource object to walk * @start: Start of the range * @size: Size of the range * @cur: cursor object to initialize * * Start walking over the range of allocations between @start and @size. */ static inline void xe_res_first(struct ttm_resource *res, u64 start, u64 size, struct xe_res_cursor *cur) { cur->sgl = NULL; cur->dma_addr = NULL; if (!res) goto fallback; XE_WARN_ON(start + size > res->size); cur->mem_type = res->mem_type; switch (cur->mem_type) { case XE_PL_STOLEN: case XE_PL_VRAM0: case XE_PL_VRAM1: { struct drm_buddy_block *block; struct list_head *head, *next; struct drm_buddy *mm = xe_res_get_buddy(res); head = &to_xe_ttm_vram_mgr_resource(res)->blocks; block = list_first_entry_or_null(head, struct drm_buddy_block, link); if (!block) goto fallback; while (start >= drm_buddy_block_size(mm, block)) { start -= drm_buddy_block_size(mm, block); next = block->link.next; if (next != head) block = list_entry(next, struct drm_buddy_block, link); } cur->mm = mm; cur->start = drm_buddy_block_offset(block) + start; cur->size = min(drm_buddy_block_size(mm, block) - start, size); cur->remaining = size; cur->node = block; break; } default: goto fallback; } return; fallback: cur->start = start; cur->size = size; cur->remaining = size; cur->node = NULL; cur->mem_type = XE_PL_TT; XE_WARN_ON(res && start + size > res->size); } static inline void __xe_res_sg_next(struct xe_res_cursor *cur) { struct scatterlist *sgl = cur->sgl; u64 start = cur->start; while (start >= sg_dma_len(sgl)) { start -= sg_dma_len(sgl); sgl = sg_next(sgl); XE_WARN_ON(!sgl); } cur->start = start; cur->size = sg_dma_len(sgl) - start; cur->sgl = sgl; } /** * __xe_res_dma_next() - Advance the cursor when end-of-segment is reached * @cur: The cursor */ static inline void __xe_res_dma_next(struct xe_res_cursor *cur) { const struct drm_pagemap_device_addr *addr = cur->dma_addr; u64 start = cur->start; while (start >= cur->dma_seg_size) { start -= cur->dma_seg_size; addr++; cur->dma_seg_size = PAGE_SIZE << addr->order; } cur->dma_start = addr->addr; /* Coalesce array_elements */ while (cur->dma_seg_size - start < cur->remaining) { if (cur->dma_start + cur->dma_seg_size != addr[1].addr || addr->proto != addr[1].proto) break; addr++; cur->dma_seg_size += PAGE_SIZE << addr->order; } cur->dma_addr = addr; cur->start = start; cur->size = cur->dma_seg_size - start; } /** * xe_res_first_sg - initialize a xe_res_cursor with a scatter gather table * * @sg: scatter gather table to walk * @start: Start of the range * @size: Size of the range * @cur: cursor object to initialize * * Start walking over the range of allocations between @start and @size. */ static inline void xe_res_first_sg(const struct sg_table *sg, u64 start, u64 size, struct xe_res_cursor *cur) { XE_WARN_ON(!sg); cur->node = NULL; cur->start = start; cur->remaining = size; cur->size = 0; cur->dma_addr = NULL; cur->sgl = sg->sgl; cur->mem_type = XE_PL_TT; __xe_res_sg_next(cur); } /** * xe_res_first_dma - initialize a xe_res_cursor with dma_addr array * * @dma_addr: struct drm_pagemap_device_addr array to walk * @start: Start of the range * @size: Size of the range * @cur: cursor object to initialize * * Start walking over the range of allocations between @start and @size. */ static inline void xe_res_first_dma(const struct drm_pagemap_device_addr *dma_addr, u64 start, u64 size, struct xe_res_cursor *cur) { XE_WARN_ON(!dma_addr); XE_WARN_ON(!IS_ALIGNED(start, PAGE_SIZE) || !IS_ALIGNED(size, PAGE_SIZE)); cur->node = NULL; cur->start = start; cur->remaining = size; cur->dma_seg_size = PAGE_SIZE << dma_addr->order; cur->dma_start = 0; cur->size = 0; cur->dma_addr = dma_addr; __xe_res_dma_next(cur); cur->sgl = NULL; cur->mem_type = XE_PL_TT; } /** * xe_res_next - advance the cursor * * @cur: the cursor to advance * @size: number of bytes to move forward * * Move the cursor @size bytes forwrad, walking to the next node if necessary. */ static inline void xe_res_next(struct xe_res_cursor *cur, u64 size) { struct drm_buddy_block *block; struct list_head *next; u64 start; XE_WARN_ON(size > cur->remaining); cur->remaining -= size; if (!cur->remaining) return; if (cur->size > size) { cur->size -= size; cur->start += size; return; } if (cur->dma_addr) { cur->start += size; __xe_res_dma_next(cur); return; } if (cur->sgl) { cur->start += size; __xe_res_sg_next(cur); return; } switch (cur->mem_type) { case XE_PL_STOLEN: case XE_PL_VRAM0: case XE_PL_VRAM1: start = size - cur->size; block = cur->node; next = block->link.next; block = list_entry(next, struct drm_buddy_block, link); while (start >= drm_buddy_block_size(cur->mm, block)) { start -= drm_buddy_block_size(cur->mm, block); next = block->link.next; block = list_entry(next, struct drm_buddy_block, link); } cur->start = drm_buddy_block_offset(block) + start; cur->size = min(drm_buddy_block_size(cur->mm, block) - start, cur->remaining); cur->node = block; break; default: return; } } /** * xe_res_dma - return dma address of cursor at current position * * @cur: the cursor to return the dma address from */ static inline u64 xe_res_dma(const struct xe_res_cursor *cur) { if (cur->dma_addr) return cur->dma_start + cur->start; else if (cur->sgl) return sg_dma_address(cur->sgl) + cur->start; else return cur->start; } /** * xe_res_is_vram() - Whether the cursor current dma address points to * same-device VRAM * @cur: The cursor. * * Return: true iff the address returned by xe_res_dma() points to internal vram. */ static inline bool xe_res_is_vram(const struct xe_res_cursor *cur) { if (cur->dma_addr) return cur->dma_addr->proto == XE_INTERCONNECT_VRAM; switch (cur->mem_type) { case XE_PL_STOLEN: case XE_PL_VRAM0: case XE_PL_VRAM1: return true; default: break; } return false; } #endif