/* SPDX-License-Identifier: GPL-2.0-only */ #ifndef _REGION_H_ #define _REGION_H_ #include #include #include #include #include /* * Region support. * * Regions are intended to abstract away the access mechanisms for blocks of * data. This could be SPI, eMMC, or a memory region as the backing store. * They are accessed through a region_device. Subregions can be made by * chaining together multiple region_devices. */ struct region_device; /* * Returns NULL on error otherwise a buffer is returned with the contents of * the requested data at offset of size. */ void *rdev_mmap(const struct region_device *rd, size_t offset, size_t size); /* Unmap a previously mapped area. Returns 0 on success, < 0 on error. */ int rdev_munmap(const struct region_device *rd, void *mapping); /* * Returns < 0 on error otherwise returns size of data read at provided * offset filling in the buffer passed. */ ssize_t rdev_readat(const struct region_device *rd, void *b, size_t offset, size_t size); /* * Returns < 0 on error otherwise returns size of data wrote at provided * offset from the buffer passed. */ ssize_t rdev_writeat(const struct region_device *rd, const void *b, size_t offset, size_t size); /* * Returns < 0 on error otherwise returns size of data erased. * If eraseat ops is not defined it returns size which indicates * that operation was successful. */ ssize_t rdev_eraseat(const struct region_device *rd, size_t offset, size_t size); /**************************************** * Implementation of a region device * ****************************************/ /* * Create a child region of the parent provided the sub-region is within * the parent's region. Returns < 0 on error otherwise 0 on success. Note * that the child device only calls through the parent's operations. */ int rdev_chain(struct region_device *child, const struct region_device *parent, size_t offset, size_t size); /* A region_device operations. */ struct region_device_ops { void *(*mmap)(const struct region_device *, size_t, size_t); int (*munmap)(const struct region_device *, void *); ssize_t (*readat)(const struct region_device *, void *, size_t, size_t); ssize_t (*writeat)(const struct region_device *, const void *, size_t, size_t); ssize_t (*eraseat)(const struct region_device *, size_t, size_t); }; struct region { size_t offset; size_t size; }; struct region_device { const struct region_device *root; const struct region_device_ops *ops; struct region region; }; #define REGION_DEV_INIT(ops_, offset_, size_) \ { \ .root = NULL, \ .ops = (ops_), \ .region = { \ .offset = (offset_), \ .size = (size_), \ }, \ } /* Helper to dynamically initialize region device. */ void region_device_init(struct region_device *rdev, const struct region_device_ops *ops, size_t offset, size_t size); /* Return 1 if child is subregion of parent, else 0. */ int region_is_subregion(const struct region *p, const struct region *c); static inline size_t region_offset(const struct region *r) { return r->offset; } static inline size_t region_sz(const struct region *r) { return r->size; } static inline size_t region_end(const struct region *r) { return region_offset(r) + region_sz(r); } static inline bool region_overlap(const struct region *r1, const struct region *r2) { return (region_end(r1) > region_offset(r2)) && (region_offset(r1) < region_end(r2)); } static inline const struct region *region_device_region( const struct region_device *rdev) { return &rdev->region; } static inline size_t region_device_sz(const struct region_device *rdev) { return region_sz(region_device_region(rdev)); } static inline size_t region_device_offset(const struct region_device *rdev) { return region_offset(region_device_region(rdev)); } static inline size_t region_device_end(const struct region_device *rdev) { return region_end(region_device_region(rdev)); } /* Memory map entire region device. Same semantics as rdev_mmap() above. */ static inline void *rdev_mmap_full(const struct region_device *rd) { return rdev_mmap(rd, 0, region_device_sz(rd)); } static inline int rdev_chain_full(struct region_device *child, const struct region_device *parent) { /* Chain full size of parent. */ return rdev_chain(child, parent, 0, region_device_sz(parent)); } /* * Returns < 0 on error otherwise returns size of data read at provided * offset filling in the buffer passed. * * You must ensure the buffer is large enough to hold the full region_device. */ static inline ssize_t rdev_read_full(const struct region_device *rd, void *b) { return rdev_readat(rd, b, 0, region_device_sz(rd)); } /* * Compute relative offset of the child (c) w.r.t. the parent (p). Returns < 0 * when child is not within the parent's region. */ ssize_t rdev_relative_offset(const struct region_device *p, const struct region_device *c); /* Helper functions to create an rdev that represents memory. */ int rdev_chain_mem(struct region_device *child, const void *base, size_t size); int rdev_chain_mem_rw(struct region_device *child, void *base, size_t size); struct mem_region_device { char *base; struct region_device rdev; }; /* Initialize at runtime a mem_region_device. Should only be used for mappings that need to fit right up to the edge of the physical address space. Most use cases will want to use rdev_chain_mem() instead. */ void mem_region_device_ro_init(struct mem_region_device *mdev, void *base, size_t size); void mem_region_device_rw_init(struct mem_region_device *mdev, void *base, size_t size); extern const struct region_device_ops mem_rdev_ro_ops; extern const struct region_device_ops mem_rdev_rw_ops; /* Statically initialize mem_region_device. Should normally only be used for const globals. Most use cases will want to use rdev_chain_mem() instead. */ #define MEM_REGION_DEV_INIT(base_, size_, ops_) \ { \ .base = (void *)(base_), \ .rdev = REGION_DEV_INIT((ops_), 0, (size_)), \ } #define MEM_REGION_DEV_RO_INIT(base_, size_) \ MEM_REGION_DEV_INIT(base_, size_, &mem_rdev_ro_ops) \ #define MEM_REGION_DEV_RW_INIT(base_, size_) \ MEM_REGION_DEV_INIT(base_, size_, &mem_rdev_rw_ops) \ struct mmap_helper_region_device { struct mem_pool *pool; struct region_device rdev; }; #define MMAP_HELPER_DEV_INIT(ops_, offset_, size_, mpool_) \ { \ .rdev = REGION_DEV_INIT((ops_), (offset_), (size_)), \ .pool = (mpool_), \ } void *mmap_helper_rdev_mmap(const struct region_device *, size_t, size_t); int mmap_helper_rdev_munmap(const struct region_device *, void *); /* * A translated region device provides the ability to publish a region device in one address * space and use an access mechanism within another address space. The sub region is the window * within the 1st address space and the request is modified prior to accessing the second * address space provided by access_dev. * * Each xlate_region_device can support multiple translation windows described using * xlate_window structure. The windows need not be contiguous in either address space. However, * this poses restrictions on the operations being performed i.e. callers cannot perform * operations across multiple windows of a translated region device. It is possible to support * readat/writeat/eraseat by translating them into multiple calls one to access device in each * window. However, mmap support is tricky because the caller expects that the memory mapped * region is contiguous in both address spaces. Thus, to keep the semantics consistent for all * region ops, xlate_region_device does not support any operations across the window * boundary. * * Note: The platform is expected to ensure that the fmap description does not place any * section (that will be operated using the translated region device) across multiple windows. */ struct xlate_window { const struct region_device *access_dev; struct region sub_region; }; struct xlate_region_device { size_t window_count; const struct xlate_window *window_arr; struct region_device rdev; }; extern const struct region_device_ops xlate_rdev_ro_ops; extern const struct region_device_ops xlate_rdev_rw_ops; #define XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, ops_) \ { \ .window_count = ARRAY_SIZE(window_arr_), \ .window_arr = window_arr_, \ .rdev = REGION_DEV_INIT((ops_), 0, (parent_sz_)), \ } #define XLATE_REGION_DEV_RO_INIT(window_arr_, parent_sz_) \ XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, &xlate_rdev_ro_ops) #define XLATE_REGION_DEV_RW_INIT(window_count_, window_arr_, parent_sz_) \ XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, &xlate_rdev_rw_ops) /* Helper to dynamically initialize xlate region device. */ void xlate_region_device_ro_init(struct xlate_region_device *xdev, size_t window_count, const struct xlate_window *window_arr, size_t parent_size); void xlate_region_device_rw_init(struct xlate_region_device *xdev, size_t window_count, const struct xlate_window *window_arr, size_t parent_size); void xlate_window_init(struct xlate_window *window, const struct region_device *access_dev, size_t sub_region_offset, size_t sub_region_size); /* This type can be used for incoherent access where the read and write * operations are backed by separate drivers. An example is x86 systems * with memory mapped media for reading but use a spi flash driver for * writing. One needs to ensure using this object is appropriate in context. */ struct incoherent_rdev { struct region_device rdev; const struct region_device *read; const struct region_device *write; }; /* Initialize an incoherent_rdev based on the region as well as the read and * write rdevs. The read and write rdevs should match in size to the passed * in region. If not the initialization will fail returning NULL. Otherwise * the function will return a pointer to the containing region_device to * be used for region operations. Therefore, the lifetime of the returned * pointer matches the lifetime of the incoherent_rdev object. Likewise, * the lifetime of the read and write rdev need to match the lifetime of * the incoherent_rdev object. */ const struct region_device *incoherent_rdev_init(struct incoherent_rdev *irdev, const struct region *r, const struct region_device *read, const struct region_device *write); #endif /* _REGION_H_ */