mm.c (7403B)
1/* 2 * Copyright 2012 Red Hat Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 * Authors: Ben Skeggs 23 */ 24#include <core/mm.h> 25 26#define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ 27 list_entry((root)->nl_entry.dir, struct nvkm_mm_node, nl_entry) 28 29void 30nvkm_mm_dump(struct nvkm_mm *mm, const char *header) 31{ 32 struct nvkm_mm_node *node; 33 34 pr_err("nvkm: %s\n", header); 35 pr_err("nvkm: node list:\n"); 36 list_for_each_entry(node, &mm->nodes, nl_entry) { 37 pr_err("nvkm: \t%08x %08x %d\n", 38 node->offset, node->length, node->type); 39 } 40 pr_err("nvkm: free list:\n"); 41 list_for_each_entry(node, &mm->free, fl_entry) { 42 pr_err("nvkm: \t%08x %08x %d\n", 43 node->offset, node->length, node->type); 44 } 45} 46 47void 48nvkm_mm_free(struct nvkm_mm *mm, struct nvkm_mm_node **pthis) 49{ 50 struct nvkm_mm_node *this = *pthis; 51 52 if (this) { 53 struct nvkm_mm_node *prev = node(this, prev); 54 struct nvkm_mm_node *next = node(this, next); 55 56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { 57 prev->length += this->length; 58 list_del(&this->nl_entry); 59 kfree(this); this = prev; 60 } 61 62 if (next && next->type == NVKM_MM_TYPE_NONE) { 63 next->offset = this->offset; 64 next->length += this->length; 65 if (this->type == NVKM_MM_TYPE_NONE) 66 list_del(&this->fl_entry); 67 list_del(&this->nl_entry); 68 kfree(this); this = NULL; 69 } 70 71 if (this && this->type != NVKM_MM_TYPE_NONE) { 72 list_for_each_entry(prev, &mm->free, fl_entry) { 73 if (this->offset < prev->offset) 74 break; 75 } 76 77 list_add_tail(&this->fl_entry, &prev->fl_entry); 78 this->type = NVKM_MM_TYPE_NONE; 79 } 80 } 81 82 *pthis = NULL; 83} 84 85static struct nvkm_mm_node * 86region_head(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) 87{ 88 struct nvkm_mm_node *b; 89 90 if (a->length == size) 91 return a; 92 93 b = kmalloc(sizeof(*b), GFP_KERNEL); 94 if (unlikely(b == NULL)) 95 return NULL; 96 97 b->offset = a->offset; 98 b->length = size; 99 b->heap = a->heap; 100 b->type = a->type; 101 a->offset += size; 102 a->length -= size; 103 list_add_tail(&b->nl_entry, &a->nl_entry); 104 if (b->type == NVKM_MM_TYPE_NONE) 105 list_add_tail(&b->fl_entry, &a->fl_entry); 106 107 return b; 108} 109 110int 111nvkm_mm_head(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, 112 u32 align, struct nvkm_mm_node **pnode) 113{ 114 struct nvkm_mm_node *prev, *this, *next; 115 u32 mask = align - 1; 116 u32 splitoff; 117 u32 s, e; 118 119 BUG_ON(type == NVKM_MM_TYPE_NONE || type == NVKM_MM_TYPE_HOLE); 120 121 list_for_each_entry(this, &mm->free, fl_entry) { 122 if (unlikely(heap != NVKM_MM_HEAP_ANY)) { 123 if (this->heap != heap) 124 continue; 125 } 126 e = this->offset + this->length; 127 s = this->offset; 128 129 prev = node(this, prev); 130 if (prev && prev->type != type) 131 s = roundup(s, mm->block_size); 132 133 next = node(this, next); 134 if (next && next->type != type) 135 e = rounddown(e, mm->block_size); 136 137 s = (s + mask) & ~mask; 138 e &= ~mask; 139 if (s > e || e - s < size_min) 140 continue; 141 142 splitoff = s - this->offset; 143 if (splitoff && !region_head(mm, this, splitoff)) 144 return -ENOMEM; 145 146 this = region_head(mm, this, min(size_max, e - s)); 147 if (!this) 148 return -ENOMEM; 149 150 this->next = NULL; 151 this->type = type; 152 list_del(&this->fl_entry); 153 *pnode = this; 154 return 0; 155 } 156 157 return -ENOSPC; 158} 159 160static struct nvkm_mm_node * 161region_tail(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) 162{ 163 struct nvkm_mm_node *b; 164 165 if (a->length == size) 166 return a; 167 168 b = kmalloc(sizeof(*b), GFP_KERNEL); 169 if (unlikely(b == NULL)) 170 return NULL; 171 172 a->length -= size; 173 b->offset = a->offset + a->length; 174 b->length = size; 175 b->heap = a->heap; 176 b->type = a->type; 177 178 list_add(&b->nl_entry, &a->nl_entry); 179 if (b->type == NVKM_MM_TYPE_NONE) 180 list_add(&b->fl_entry, &a->fl_entry); 181 182 return b; 183} 184 185int 186nvkm_mm_tail(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, 187 u32 align, struct nvkm_mm_node **pnode) 188{ 189 struct nvkm_mm_node *prev, *this, *next; 190 u32 mask = align - 1; 191 192 BUG_ON(type == NVKM_MM_TYPE_NONE || type == NVKM_MM_TYPE_HOLE); 193 194 list_for_each_entry_reverse(this, &mm->free, fl_entry) { 195 u32 e = this->offset + this->length; 196 u32 s = this->offset; 197 u32 c = 0, a; 198 if (unlikely(heap != NVKM_MM_HEAP_ANY)) { 199 if (this->heap != heap) 200 continue; 201 } 202 203 prev = node(this, prev); 204 if (prev && prev->type != type) 205 s = roundup(s, mm->block_size); 206 207 next = node(this, next); 208 if (next && next->type != type) { 209 e = rounddown(e, mm->block_size); 210 c = next->offset - e; 211 } 212 213 s = (s + mask) & ~mask; 214 a = e - s; 215 if (s > e || a < size_min) 216 continue; 217 218 a = min(a, size_max); 219 s = (e - a) & ~mask; 220 c += (e - s) - a; 221 222 if (c && !region_tail(mm, this, c)) 223 return -ENOMEM; 224 225 this = region_tail(mm, this, a); 226 if (!this) 227 return -ENOMEM; 228 229 this->next = NULL; 230 this->type = type; 231 list_del(&this->fl_entry); 232 *pnode = this; 233 return 0; 234 } 235 236 return -ENOSPC; 237} 238 239int 240nvkm_mm_init(struct nvkm_mm *mm, u8 heap, u32 offset, u32 length, u32 block) 241{ 242 struct nvkm_mm_node *node, *prev; 243 u32 next; 244 245 if (nvkm_mm_initialised(mm)) { 246 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); 247 next = prev->offset + prev->length; 248 if (next != offset) { 249 BUG_ON(next > offset); 250 if (!(node = kzalloc(sizeof(*node), GFP_KERNEL))) 251 return -ENOMEM; 252 node->type = NVKM_MM_TYPE_HOLE; 253 node->offset = next; 254 node->length = offset - next; 255 list_add_tail(&node->nl_entry, &mm->nodes); 256 } 257 BUG_ON(block != mm->block_size); 258 } else { 259 INIT_LIST_HEAD(&mm->nodes); 260 INIT_LIST_HEAD(&mm->free); 261 mm->block_size = block; 262 mm->heap_nodes = 0; 263 } 264 265 node = kzalloc(sizeof(*node), GFP_KERNEL); 266 if (!node) 267 return -ENOMEM; 268 269 if (length) { 270 node->offset = roundup(offset, mm->block_size); 271 node->length = rounddown(offset + length, mm->block_size); 272 node->length -= node->offset; 273 } 274 275 list_add_tail(&node->nl_entry, &mm->nodes); 276 list_add_tail(&node->fl_entry, &mm->free); 277 node->heap = heap; 278 mm->heap_nodes++; 279 return 0; 280} 281 282int 283nvkm_mm_fini(struct nvkm_mm *mm) 284{ 285 struct nvkm_mm_node *node, *temp; 286 int nodes = 0; 287 288 if (!nvkm_mm_initialised(mm)) 289 return 0; 290 291 list_for_each_entry(node, &mm->nodes, nl_entry) { 292 if (node->type != NVKM_MM_TYPE_HOLE) { 293 if (++nodes > mm->heap_nodes) { 294 nvkm_mm_dump(mm, "mm not clean!"); 295 return -EBUSY; 296 } 297 } 298 } 299 300 list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) { 301 list_del(&node->nl_entry); 302 kfree(node); 303 } 304 305 mm->heap_nodes = 0; 306 return 0; 307}