Allow for resizing of LMB regions if the region attributes match. The
current code returns a failure status on detecting an overlapping
address. This worked up until now since the LMB calls were not
persistent and global -- the LMB memory map was specific and private
to a given caller of the LMB API's.
With the change in the LMB code to make the LMB reservations
persistent, there needs to be a check on whether the memory region can
be resized, and then do it if so. To distinguish between memory that
cannot be resized, add a new flag, LMB_NOOVERWRITE. Reserving a region
of memory with this attribute would indicate that the region cannot be
resized.

Signed-off-by: Sughosh Ganu <sughosh.g...@linaro.org>
---
Changes since rfc:
* Add a function comment for lmb_add_region_flags().
* Change the wording of a comment in lmb_merge_overlap_regions() as
  per review comment from Simon Glass.

 include/lmb.h |   1 +
 lib/lmb.c     | 144 ++++++++++++++++++++++++++++++++++++++++++++------
 2 files changed, 128 insertions(+), 17 deletions(-)

diff --git a/include/lmb.h b/include/lmb.h
index a1cc45b726..a308796d58 100644
--- a/include/lmb.h
+++ b/include/lmb.h
@@ -21,6 +21,7 @@
 enum lmb_flags {
        LMB_NONE                = BIT(0),
        LMB_NOMAP               = BIT(1),
+       LMB_NOOVERWRITE         = BIT(2),
 };
 
 /**
diff --git a/lib/lmb.c b/lib/lmb.c
index dd6f22654c..88352e9a25 100644
--- a/lib/lmb.c
+++ b/lib/lmb.c
@@ -247,12 +247,106 @@ void lmb_init_and_reserve_range(phys_addr_t base, 
phys_size_t size,
        lmb_reserve_common(fdt_blob);
 }
 
-/* This routine called with relocation disabled. */
+static bool lmb_region_flags_match(struct lmb_region *rgn, unsigned long r1,
+                                  enum lmb_flags flags)
+{
+       return rgn[r1].flags == flags;
+}
+
+static long lmb_merge_overlap_regions(struct alist *lmb_rgn_lst,
+                                     unsigned long i, phys_addr_t base,
+                                     phys_size_t size, enum lmb_flags flags)
+{
+       phys_size_t rgnsize;
+       unsigned long rgn_cnt, idx;
+       phys_addr_t rgnbase, rgnend;
+       phys_addr_t mergebase, mergeend;
+       struct lmb_region *rgn = lmb_rgn_lst->data;
+
+       rgn_cnt = 0;
+       idx = i;
+
+       /*
+        * First thing to do is to identify how many regions
+        * the requested region overlaps.
+        * If the flags match, combine all these overlapping
+        * regions into a single region, and remove the merged
+        * regions.
+        */
+       while (idx < lmb_rgn_lst->count - 1) {
+               rgnbase = rgn[idx].base;
+               rgnsize = rgn[idx].size;
+
+               if (lmb_addrs_overlap(base, size, rgnbase,
+                                     rgnsize)) {
+                       if (!lmb_region_flags_match(rgn, idx, flags))
+                               return -1;
+                       rgn_cnt++;
+                       idx++;
+               }
+       }
+
+       /* The merged region's base and size */
+       rgnbase = rgn[i].base;
+       mergebase = min(base, rgnbase);
+       rgnend = rgn[idx].base + rgn[idx].size;
+       mergeend = max(rgnend, (base + size));
+
+       rgn[i].base = mergebase;
+       rgn[i].size = mergeend - mergebase;
+
+       /* Now remove the merged regions */
+       while (--rgn_cnt)
+               lmb_remove_region(lmb_rgn_lst, i + 1);
+
+       return 0;
+}
+
+static long lmb_resize_regions(struct alist *lmb_rgn_lst, unsigned long i,
+                              phys_addr_t base, phys_size_t size,
+                              enum lmb_flags flags)
+{
+       long ret = 0;
+       phys_addr_t rgnend;
+       struct lmb_region *rgn = lmb_rgn_lst->data;
+
+       if (i == lmb_rgn_lst->count - 1 ||
+               base + size < rgn[i + 1].base) {
+               if (!lmb_region_flags_match(rgn, i, flags))
+                       return -1;
+
+               rgnend = rgn[i].base + rgn[i].size;
+               rgn[i].base = min(base, rgn[i].base);
+               rgnend = max(base + size, rgnend);
+               rgn[i].size = rgnend - rgn[i].base;
+       } else {
+               ret = lmb_merge_overlap_regions(lmb_rgn_lst, i, base, size,
+                                               flags);
+       }
+
+       return ret;
+}
+
+/**
+ * lmb_add_region_flags() - Add an lmb region to the given list
+ * @lmb_rgn_lst: LMB list to which region is to be added(free/used)
+ * @base: Start address of the region
+ * @size: Size of the region to be added
+ * @flags: Attributes of the LMB region
+ *
+ * Add a region of memory to the list. If the region does not exist, add
+ * it to the list. Depending on the attributes of the region to be added,
+ * the function might resize an already existing region or coalesce two
+ * adjacent regions.
+ *
+ *
+ * Returns: 0 if the region addition successful, -1 on failure
+ */
 static long lmb_add_region_flags(struct alist *lmb_rgn_lst, phys_addr_t base,
                                 phys_size_t size, enum lmb_flags flags)
 {
        unsigned long coalesced = 0;
-       long adjacent, i;
+       long ret, i;
        struct lmb_region *rgn = lmb_rgn_lst->data;
 
        if (alist_err(lmb_rgn_lst))
@@ -281,23 +375,32 @@ static long lmb_add_region_flags(struct alist 
*lmb_rgn_lst, phys_addr_t base,
                                return -1; /* regions with new flags */
                }
 
-               adjacent = lmb_addrs_adjacent(base, size, rgnbase, rgnsize);
-               if (adjacent > 0) {
+               ret = lmb_addrs_adjacent(base, size, rgnbase, rgnsize);
+               if (ret > 0) {
                        if (flags != rgnflags)
                                break;
                        rgn[i].base -= size;
                        rgn[i].size += size;
                        coalesced++;
                        break;
-               } else if (adjacent < 0) {
+               } else if (ret < 0) {
                        if (flags != rgnflags)
                                break;
                        rgn[i].size += size;
                        coalesced++;
                        break;
                } else if (lmb_addrs_overlap(base, size, rgnbase, rgnsize)) {
-                       /* regions overlap */
-                       return -1;
+                       if (flags == LMB_NONE) {
+                               ret = lmb_resize_regions(lmb_rgn_lst, i, base,
+                                                        size, flags);
+                               if (ret < 0)
+                                       return -1;
+
+                               coalesced++;
+                               break;
+                       } else {
+                               return -1;
+                       }
                }
        }
 
@@ -447,7 +550,7 @@ static phys_addr_t lmb_align_down(phys_addr_t addr, 
phys_size_t size)
 }
 
 static phys_addr_t __lmb_alloc_base(phys_size_t size, ulong align,
-                                   phys_addr_t max_addr)
+                                   phys_addr_t max_addr, enum lmb_flags flags)
 {
        long i, rgn;
        phys_addr_t base = 0;
@@ -476,8 +579,8 @@ static phys_addr_t __lmb_alloc_base(phys_size_t size, ulong 
align,
                        rgn = lmb_overlaps_region(&lmb_used_mem, base, size);
                        if (rgn < 0) {
                                /* This area isn't reserved, take it */
-                               if (lmb_add_region(&lmb_used_mem, base,
-                                                  size) < 0)
+                               if (lmb_add_region_flags(&lmb_used_mem, base,
+                                                        size, flags) < 0)
                                        return 0;
                                return base;
                        }
@@ -500,7 +603,7 @@ phys_addr_t lmb_alloc_base(phys_size_t size, ulong align, 
phys_addr_t max_addr)
 {
        phys_addr_t alloc;
 
-       alloc = __lmb_alloc_base(size, align, max_addr);
+       alloc = __lmb_alloc_base(size, align, max_addr, LMB_NONE);
 
        if (alloc == 0)
                printf("ERROR: Failed to allocate 0x%lx bytes below 0x%lx.\n",
@@ -509,11 +612,8 @@ phys_addr_t lmb_alloc_base(phys_size_t size, ulong align, 
phys_addr_t max_addr)
        return alloc;
 }
 
-/*
- * Try to allocate a specific address range: must be in defined memory but not
- * reserved
- */
-phys_addr_t lmb_alloc_addr(phys_addr_t base, phys_size_t size)
+static phys_addr_t __lmb_alloc_addr(phys_addr_t base, phys_size_t size,
+                                   enum lmb_flags flags)
 {
        long rgn;
        struct lmb_region *lmb_memory = lmb_free_mem.data;
@@ -529,13 +629,23 @@ phys_addr_t lmb_alloc_addr(phys_addr_t base, phys_size_t 
size)
                                      lmb_memory[rgn].size,
                                      base + size - 1, 1)) {
                        /* ok, reserve the memory */
-                       if (lmb_reserve(base, size) >= 0)
+                       if (lmb_reserve_flags(base, size, flags) >= 0)
                                return base;
                }
        }
+
        return 0;
 }
 
+/*
+ * Try to allocate a specific address range: must be in defined memory but not
+ * reserved
+ */
+phys_addr_t lmb_alloc_addr(phys_addr_t base, phys_size_t size)
+{
+       return __lmb_alloc_addr(base, size, LMB_NONE);
+}
+
 /* Return number of bytes from a given address that are free */
 phys_size_t lmb_get_free_size(phys_addr_t addr)
 {
-- 
2.34.1

Reply via email to