/*- * Copyright (c) 2005-2006, Kohsuke Ohtani * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the author nor the names of any co-contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ /* * page.c - physical page allocator */ /* * This is a simple list-based page allocator. * * When the remaining page is exhausted, what should we do ? * If the system can stop with panic() here, the error check of many * portions in kernel is not necessary, and kernel code can become * more simple. But, in general, even if a page is exhausted, * a kernel can not be stopped but it should return an error and * continue processing. * If the memory becomes short during boot time, kernel and drivers * can use panic() in that case. */ #include #include #include /* * page_block is put on the head of the first page of * each free block. */ struct page_block { struct page_block *next; struct page_block *prev; size_t size; /* number of bytes of this block */ }; static struct page_block page_head; /* first free block */ static size_t total_bytes; static size_t used_bytes; /* * page_alloc - allocate continuous pages of the specified size. * * This routine returns the physical address of a new free page * block, or returns NULL on failure. The requested size is * automatically round up to the page boundary. The allocated * memory is _not_ filled with 0. */ void * page_alloc(size_t size) { struct page_block *blk, *tmp; ASSERT(size != 0); sched_lock(); /* * Find the free block that has enough size. */ size = (size_t)PAGE_ALIGN(size); blk = &page_head; do { blk = blk->next; if (blk == &page_head) { sched_unlock(); DPRINTF(("page_alloc: out of memory\n")); return NULL; /* Not found. */ } } while (blk->size < size); /* * If found block size is exactly same with requested, * just remove it from a free list. Otherwise, the * found block is divided into two and first half is * used for allocation. */ if (blk->size == size) { blk->prev->next = blk->next; blk->next->prev = blk->prev; } else { tmp = (struct page_block *)((char *)blk + size); tmp->size = blk->size - size; tmp->prev = blk->prev; tmp->next = blk->next; blk->prev->next = tmp; blk->next->prev = tmp; } used_bytes += size; sched_unlock(); return virt_to_phys(blk); } /* * Free page block. * * This allocator does not maintain the size of allocated page * block. The caller must provide the size information of the * block. */ void page_free(void *addr, size_t size) { struct page_block *blk, *prev; ASSERT(addr != NULL); ASSERT(size != 0); sched_lock(); size = (size_t)PAGE_ALIGN(size); blk = (struct page_block *)phys_to_virt(addr); /* * Find the target position in list. */ for (prev = &page_head; prev->next < blk; prev = prev->next) { if (prev->next == &page_head) break; } #ifdef DEBUG if (prev != &page_head) ASSERT((char *)prev + prev->size <= (char *)blk); if (prev->next != &page_head) ASSERT((char *)blk + size <= (char *)prev->next); #endif /* DEBUG */ /* * Insert new block into list. */ blk->size = size; blk->prev = prev; blk->next = prev->next; prev->next->prev = blk; prev->next = blk; /* * If the adjoining block is free, it combines and * is made on block. */ if (blk->next != &page_head && ((char *)blk + blk->size) == (char *)blk->next) { blk->size += blk->next->size; blk->next = blk->next->next; blk->next->prev = blk; } if (blk->prev != &page_head && (char *)blk->prev + blk->prev->size == (char *)blk) { blk->prev->size += blk->size; blk->prev->next = blk->next; blk->next->prev = blk->prev; } used_bytes -= size; sched_unlock(); } /* * The function to reserve pages in specific address. * Return 0 on success, or -1 on failure */ int page_reserve(void *addr, size_t size) { struct page_block *blk, *tmp; char *end; if (size == 0) return 0; addr = phys_to_virt(addr); end = (char *)PAGE_ALIGN((char *)addr + size); addr = (void *)PAGE_TRUNC(addr); size = (size_t)(end - (char *)addr); /* * Find the block which includes specified block. */ blk = page_head.next; for (;;) { if (blk == &page_head) panic("failed to reserve pages"); if ((char *)blk <= (char *)addr && end <= (char *)blk + blk->size) break; blk = blk->next; } if ((char *)blk == (char *)addr && blk->size == size) { /* * Unlink the block from free list. */ blk->prev->next = blk->next; blk->next->prev = blk->prev; } else { /* * Split this block. */ if ((char *)blk + blk->size != end) { tmp = (struct page_block *)end; tmp->size = (size_t)((char *)blk + blk->size - end); tmp->next = blk->next; tmp->prev = blk; blk->size -= tmp->size; blk->next->prev = tmp; blk->next = tmp; } if ((char *)blk == (char *)addr) { blk->prev->next = blk->next; blk->next->prev = blk->prev; } else blk->size = (size_t)((char *)addr - (char *)blk); } used_bytes += size; return 0; } void page_info(size_t *total, size_t *free) { *total = total_bytes; *free = total_bytes - used_bytes; } /* * Initialize page allocator. * page_init() must be called prior to other memory manager's * initializations. */ void page_init(void) { struct page_block *blk; struct mem_map *mem; int i; DPRINTF(("Memory: base=%x size=%dK\n", boot_info->main_mem.start, boot_info->main_mem.size / 1024)); /* * First, create one block containing all memory pages. */ blk = (struct page_block *)boot_info->main_mem.start; blk = phys_to_virt(blk); blk->size = boot_info->main_mem.size; if (blk->size == 0) panic("page_init: no pages"); blk->prev = blk->next = &page_head; page_head.next = page_head.prev = blk; /* * Then, the system reserved pages are marked as a used block. */ for (i = 0; i < NRESMEM; i++) { mem = &boot_info->reserved[i]; if (mem->size != 0) page_reserve((void *)mem->start, mem->size); } total_bytes = boot_info->main_mem.size - used_bytes; used_bytes = 0; /* * Reserve pages for all boot modules. */ mem = &boot_info->modules; page_reserve((void *)mem->start, mem->size); }