summaryrefslogtreecommitdiff
path: root/vm/vm_page.c
diff options
context:
space:
mode:
Diffstat (limited to 'vm/vm_page.c')
-rw-r--r--vm/vm_page.c331
1 files changed, 166 insertions, 165 deletions
diff --git a/vm/vm_page.c b/vm/vm_page.c
index daad01fa..19f74a40 100644
--- a/vm/vm_page.c
+++ b/vm/vm_page.c
@@ -51,13 +51,13 @@
#define DEBUG 0
/*
- * Number of free block lists per segment.
+ * Number of free block lists per zone.
*/
#define VM_PAGE_NR_FREE_LISTS 11
/*
* The size of a CPU pool is computed by dividing the number of pages in its
- * containing segment by this value.
+ * containing zone by this value.
*/
#define VM_PAGE_CPU_POOL_RATIO 1024
@@ -98,14 +98,14 @@ struct vm_page_free_list {
};
/*
- * Segment name buffer size.
+ * Zone name buffer size.
*/
#define VM_PAGE_NAME_SIZE 16
/*
- * Segment of contiguous memory.
+ * Zone of contiguous memory.
*/
-struct vm_page_seg {
+struct vm_page_zone {
struct vm_page_cpu_pool cpu_pools[MAX_CPUS];
phys_addr_t start;
@@ -118,9 +118,9 @@ struct vm_page_seg {
};
/*
- * Bootstrap information about a segment.
+ * Bootstrap information about a zone.
*/
-struct vm_page_boot_seg {
+struct vm_page_boot_zone {
phys_addr_t start;
phys_addr_t end;
bool heap_present;
@@ -131,39 +131,40 @@ struct vm_page_boot_seg {
static int vm_page_is_ready __read_mostly;
/*
- * Segment table.
+ * Zone table.
*
- * The system supports a maximum of 4 segments :
+ * The system supports a maximum of 4 zones :
* - DMA: suitable for DMA
* - DMA32: suitable for DMA when devices support 32-bits addressing
* - DIRECTMAP: direct physical mapping, allows direct access from
* the kernel with a simple offset translation
* - HIGHMEM: must be mapped before it can be accessed
*
- * Segments are ordered by priority, 0 being the lowest priority. Their
- * relative priorities are DMA < DMA32 < DIRECTMAP < HIGHMEM. Some segments
+ * Zones are ordered by priority, 0 being the lowest priority. Their
+ * relative priorities are DMA < DMA32 < DIRECTMAP < HIGHMEM. Some zones
* may actually be aliases for others, e.g. if DMA is always possible from
* the direct physical mapping, DMA and DMA32 are aliases for DIRECTMAP,
- * in which case the segment table contains DIRECTMAP and HIGHMEM only.
+ * in which case the zone table contains DIRECTMAP and HIGHMEM only.
*/
-static struct vm_page_seg vm_page_segs[VM_PAGE_MAX_SEGS];
+static struct vm_page_zone vm_page_zones[VM_PAGE_MAX_ZONES];
/*
- * Bootstrap segment table.
+ * Bootstrap zone table.
*/
-static struct vm_page_boot_seg vm_page_boot_segs[VM_PAGE_MAX_SEGS] __initdata;
+static struct vm_page_boot_zone vm_page_boot_zones[VM_PAGE_MAX_ZONES]
+ __initdata;
/*
- * Number of loaded segments.
+ * Number of loaded zones.
*/
-static unsigned int vm_page_segs_size __read_mostly;
+static unsigned int vm_page_zones_size __read_mostly;
static void __init
-vm_page_init(struct vm_page *page, unsigned short seg_index, phys_addr_t pa)
+vm_page_init(struct vm_page *page, unsigned short zone_index, phys_addr_t pa)
{
memset(page, 0, sizeof(*page));
page->type = VM_PAGE_RESERVED;
- page->seg_index = seg_index;
+ page->zone_index = zone_index;
page->order = VM_PAGE_ORDER_UNLISTED;
page->phys_addr = pa;
}
@@ -208,7 +209,7 @@ vm_page_free_list_remove(struct vm_page_free_list *free_list,
}
static struct vm_page *
-vm_page_seg_alloc_from_buddy(struct vm_page_seg *seg, unsigned int order)
+vm_page_zone_alloc_from_buddy(struct vm_page_zone *zone, unsigned int order)
{
struct vm_page_free_list *free_list = free_list;
struct vm_page *page, *buddy;
@@ -217,7 +218,7 @@ vm_page_seg_alloc_from_buddy(struct vm_page_seg *seg, unsigned int order)
assert(order < VM_PAGE_NR_FREE_LISTS);
for (i = order; i < VM_PAGE_NR_FREE_LISTS; i++) {
- free_list = &seg->free_lists[i];
+ free_list = &zone->free_lists[i];
if (free_list->size != 0) {
break;
@@ -235,24 +236,24 @@ vm_page_seg_alloc_from_buddy(struct vm_page_seg *seg, unsigned int order)
while (i > order) {
i--;
buddy = &page[1 << i];
- vm_page_free_list_insert(&seg->free_lists[i], buddy);
+ vm_page_free_list_insert(&zone->free_lists[i], buddy);
buddy->order = i;
}
- seg->nr_free_pages -= (1 << order);
+ zone->nr_free_pages -= (1 << order);
return page;
}
static void
-vm_page_seg_free_to_buddy(struct vm_page_seg *seg, struct vm_page *page,
- unsigned int order)
+vm_page_zone_free_to_buddy(struct vm_page_zone *zone, struct vm_page *page,
+ unsigned int order)
{
struct vm_page *buddy;
phys_addr_t pa, buddy_pa;
unsigned int nr_pages;
- assert(page >= seg->pages);
- assert(page < seg->pages_end);
+ assert(page >= zone->pages);
+ assert(page < zone->pages_end);
assert(page->order == VM_PAGE_ORDER_UNLISTED);
assert(order < VM_PAGE_NR_FREE_LISTS);
@@ -262,26 +263,26 @@ vm_page_seg_free_to_buddy(struct vm_page_seg *seg, struct vm_page *page,
while (order < (VM_PAGE_NR_FREE_LISTS - 1)) {
buddy_pa = pa ^ vm_page_ptoa(1 << order);
- if ((buddy_pa < seg->start) || (buddy_pa >= seg->end)) {
+ if ((buddy_pa < zone->start) || (buddy_pa >= zone->end)) {
break;
}
- buddy = &seg->pages[vm_page_atop(buddy_pa - seg->start)];
+ buddy = &zone->pages[vm_page_atop(buddy_pa - zone->start)];
if (buddy->order != order) {
break;
}
- vm_page_free_list_remove(&seg->free_lists[order], buddy);
+ vm_page_free_list_remove(&zone->free_lists[order], buddy);
buddy->order = VM_PAGE_ORDER_UNLISTED;
order++;
pa &= -vm_page_ptoa(1 << order);
- page = &seg->pages[vm_page_atop(pa - seg->start)];
+ page = &zone->pages[vm_page_atop(pa - zone->start)];
}
- vm_page_free_list_insert(&seg->free_lists[order], page);
+ vm_page_free_list_insert(&zone->free_lists[order], page);
page->order = order;
- seg->nr_free_pages += nr_pages;
+ zone->nr_free_pages += nr_pages;
}
static void __init
@@ -296,9 +297,9 @@ vm_page_cpu_pool_init(struct vm_page_cpu_pool *cpu_pool, int size)
}
static inline struct vm_page_cpu_pool *
-vm_page_cpu_pool_get(struct vm_page_seg *seg)
+vm_page_cpu_pool_get(struct vm_page_zone *zone)
{
- return &seg->cpu_pools[cpu_id()];
+ return &zone->cpu_pools[cpu_id()];
}
static inline struct vm_page *
@@ -323,17 +324,17 @@ vm_page_cpu_pool_push(struct vm_page_cpu_pool *cpu_pool, struct vm_page *page)
static int
vm_page_cpu_pool_fill(struct vm_page_cpu_pool *cpu_pool,
- struct vm_page_seg *seg)
+ struct vm_page_zone *zone)
{
struct vm_page *page;
int i;
assert(cpu_pool->nr_pages == 0);
- mutex_lock(&seg->lock);
+ mutex_lock(&zone->lock);
for (i = 0; i < cpu_pool->transfer_size; i++) {
- page = vm_page_seg_alloc_from_buddy(seg, 0);
+ page = vm_page_zone_alloc_from_buddy(zone, 0);
if (page == NULL) {
break;
@@ -342,42 +343,42 @@ vm_page_cpu_pool_fill(struct vm_page_cpu_pool *cpu_pool,
vm_page_cpu_pool_push(cpu_pool, page);
}
- mutex_unlock(&seg->lock);
+ mutex_unlock(&zone->lock);
return i;
}
static void
vm_page_cpu_pool_drain(struct vm_page_cpu_pool *cpu_pool,
- struct vm_page_seg *seg)
+ struct vm_page_zone *zone)
{
struct vm_page *page;
int i;
assert(cpu_pool->nr_pages == cpu_pool->size);
- mutex_lock(&seg->lock);
+ mutex_lock(&zone->lock);
for (i = cpu_pool->transfer_size; i > 0; i--) {
page = vm_page_cpu_pool_pop(cpu_pool);
- vm_page_seg_free_to_buddy(seg, page, 0);
+ vm_page_zone_free_to_buddy(zone, page, 0);
}
- mutex_unlock(&seg->lock);
+ mutex_unlock(&zone->lock);
}
static phys_addr_t __init
-vm_page_seg_size(struct vm_page_seg *seg)
+vm_page_zone_size(struct vm_page_zone *zone)
{
- return seg->end - seg->start;
+ return zone->end - zone->start;
}
static int __init
-vm_page_seg_compute_pool_size(struct vm_page_seg *seg)
+vm_page_zone_compute_pool_size(struct vm_page_zone *zone)
{
phys_addr_t size;
- size = vm_page_atop(vm_page_seg_size(seg)) / VM_PAGE_CPU_POOL_RATIO;
+ size = vm_page_atop(vm_page_zone_size(zone)) / VM_PAGE_CPU_POOL_RATIO;
if (size == 0) {
size = 1;
@@ -389,40 +390,40 @@ vm_page_seg_compute_pool_size(struct vm_page_seg *seg)
}
static void __init
-vm_page_seg_init(struct vm_page_seg *seg, phys_addr_t start, phys_addr_t end,
- struct vm_page *pages)
+vm_page_zone_init(struct vm_page_zone *zone, phys_addr_t start, phys_addr_t end,
+ struct vm_page *pages)
{
phys_addr_t pa;
int pool_size;
unsigned int i;
- seg->start = start;
- seg->end = end;
- pool_size = vm_page_seg_compute_pool_size(seg);
+ zone->start = start;
+ zone->end = end;
+ pool_size = vm_page_zone_compute_pool_size(zone);
- for (i = 0; i < ARRAY_SIZE(seg->cpu_pools); i++) {
- vm_page_cpu_pool_init(&seg->cpu_pools[i], pool_size);
+ for (i = 0; i < ARRAY_SIZE(zone->cpu_pools); i++) {
+ vm_page_cpu_pool_init(&zone->cpu_pools[i], pool_size);
}
- seg->pages = pages;
- seg->pages_end = pages + vm_page_atop(vm_page_seg_size(seg));
- mutex_init(&seg->lock);
+ zone->pages = pages;
+ zone->pages_end = pages + vm_page_atop(vm_page_zone_size(zone));
+ mutex_init(&zone->lock);
- for (i = 0; i < ARRAY_SIZE(seg->free_lists); i++) {
- vm_page_free_list_init(&seg->free_lists[i]);
+ for (i = 0; i < ARRAY_SIZE(zone->free_lists); i++) {
+ vm_page_free_list_init(&zone->free_lists[i]);
}
- seg->nr_free_pages = 0;
- i = seg - vm_page_segs;
+ zone->nr_free_pages = 0;
+ i = zone - vm_page_zones;
- for (pa = seg->start; pa < seg->end; pa += PAGE_SIZE) {
- vm_page_init(&pages[vm_page_atop(pa - seg->start)], i, pa);
+ for (pa = zone->start; pa < zone->end; pa += PAGE_SIZE) {
+ vm_page_init(&pages[vm_page_atop(pa - zone->start)], i, pa);
}
}
static struct vm_page *
-vm_page_seg_alloc(struct vm_page_seg *seg, unsigned int order,
- unsigned short type)
+vm_page_zone_alloc(struct vm_page_zone *zone, unsigned int order,
+ unsigned short type)
{
struct vm_page_cpu_pool *cpu_pool;
struct vm_page *page;
@@ -432,11 +433,11 @@ vm_page_seg_alloc(struct vm_page_seg *seg, unsigned int order,
if (order == 0) {
thread_pin();
- cpu_pool = vm_page_cpu_pool_get(seg);
+ cpu_pool = vm_page_cpu_pool_get(zone);
mutex_lock(&cpu_pool->lock);
if (cpu_pool->nr_pages == 0) {
- filled = vm_page_cpu_pool_fill(cpu_pool, seg);
+ filled = vm_page_cpu_pool_fill(cpu_pool, zone);
if (!filled) {
mutex_unlock(&cpu_pool->lock);
@@ -449,9 +450,9 @@ vm_page_seg_alloc(struct vm_page_seg *seg, unsigned int order,
mutex_unlock(&cpu_pool->lock);
thread_unpin();
} else {
- mutex_lock(&seg->lock);
- page = vm_page_seg_alloc_from_buddy(seg, order);
- mutex_unlock(&seg->lock);
+ mutex_lock(&zone->lock);
+ page = vm_page_zone_alloc_from_buddy(zone, order);
+ mutex_unlock(&zone->lock);
if (page == NULL) {
return NULL;
@@ -464,8 +465,8 @@ vm_page_seg_alloc(struct vm_page_seg *seg, unsigned int order,
}
static void
-vm_page_seg_free(struct vm_page_seg *seg, struct vm_page *page,
- unsigned int order)
+vm_page_zone_free(struct vm_page_zone *zone, struct vm_page *page,
+ unsigned int order)
{
struct vm_page_cpu_pool *cpu_pool;
@@ -476,69 +477,69 @@ vm_page_seg_free(struct vm_page_seg *seg, struct vm_page *page,
if (order == 0) {
thread_pin();
- cpu_pool = vm_page_cpu_pool_get(seg);
+ cpu_pool = vm_page_cpu_pool_get(zone);
mutex_lock(&cpu_pool->lock);
if (cpu_pool->nr_pages == cpu_pool->size) {
- vm_page_cpu_pool_drain(cpu_pool, seg);
+ vm_page_cpu_pool_drain(cpu_pool, zone);
}
vm_page_cpu_pool_push(cpu_pool, page);
mutex_unlock(&cpu_pool->lock);
thread_unpin();
} else {
- mutex_lock(&seg->lock);
- vm_page_seg_free_to_buddy(seg, page, order);
- mutex_unlock(&seg->lock);
+ mutex_lock(&zone->lock);
+ vm_page_zone_free_to_buddy(zone, page, order);
+ mutex_unlock(&zone->lock);
}
}
void __init
-vm_page_load(unsigned int seg_index, phys_addr_t start, phys_addr_t end)
+vm_page_load(unsigned int zone_index, phys_addr_t start, phys_addr_t end)
{
- struct vm_page_boot_seg *seg;
+ struct vm_page_boot_zone *zone;
- assert(seg_index < ARRAY_SIZE(vm_page_boot_segs));
+ assert(zone_index < ARRAY_SIZE(vm_page_boot_zones));
assert(vm_page_aligned(start));
assert(vm_page_aligned(end));
assert(start < end);
- assert(vm_page_segs_size < ARRAY_SIZE(vm_page_boot_segs));
+ assert(vm_page_zones_size < ARRAY_SIZE(vm_page_boot_zones));
- seg = &vm_page_boot_segs[seg_index];
- seg->start = start;
- seg->end = end;
- seg->heap_present = false;
+ zone = &vm_page_boot_zones[zone_index];
+ zone->start = start;
+ zone->end = end;
+ zone->heap_present = false;
#if DEBUG
printk("vm_page: load: %s: %llx:%llx\n",
- vm_page_seg_name(seg_index),
+ vm_page_zone_name(zone_index),
(unsigned long long)start, (unsigned long long)end);
#endif
- vm_page_segs_size++;
+ vm_page_zones_size++;
}
void
-vm_page_load_heap(unsigned int seg_index, phys_addr_t start, phys_addr_t end)
+vm_page_load_heap(unsigned int zone_index, phys_addr_t start, phys_addr_t end)
{
- struct vm_page_boot_seg *seg;
+ struct vm_page_boot_zone *zone;
- assert(seg_index < ARRAY_SIZE(vm_page_boot_segs));
+ assert(zone_index < ARRAY_SIZE(vm_page_boot_zones));
assert(vm_page_aligned(start));
assert(vm_page_aligned(end));
- seg = &vm_page_boot_segs[seg_index];
+ zone = &vm_page_boot_zones[zone_index];
- assert(seg->start <= start);
- assert(end <= seg-> end);
+ assert(zone->start <= start);
+ assert(end <= zone-> end);
- seg->avail_start = start;
- seg->avail_end = end;
- seg->heap_present = true;
+ zone->avail_start = start;
+ zone->avail_end = end;
+ zone->heap_present = true;
#if DEBUG
printk("vm_page: heap: %s: %llx:%llx\n",
- vm_page_seg_name(seg_index),
+ vm_page_zone_name(zone_index),
(unsigned long long)start, (unsigned long long)end);
#endif
}
@@ -550,88 +551,88 @@ vm_page_ready(void)
}
static unsigned int
-vm_page_select_alloc_seg(unsigned int selector)
+vm_page_select_alloc_zone(unsigned int selector)
{
- unsigned int seg_index;
+ unsigned int zone_index;
switch (selector) {
case VM_PAGE_SEL_DMA:
- seg_index = VM_PAGE_SEG_DMA;
+ zone_index = VM_PAGE_ZONE_DMA;
break;
case VM_PAGE_SEL_DMA32:
- seg_index = VM_PAGE_SEG_DMA32;
+ zone_index = VM_PAGE_ZONE_DMA32;
break;
case VM_PAGE_SEL_DIRECTMAP:
- seg_index = VM_PAGE_SEG_DIRECTMAP;
+ zone_index = VM_PAGE_ZONE_DIRECTMAP;
break;
case VM_PAGE_SEL_HIGHMEM:
- seg_index = VM_PAGE_SEG_HIGHMEM;
+ zone_index = VM_PAGE_ZONE_HIGHMEM;
break;
default:
panic("vm_page: invalid selector");
}
- return MIN(vm_page_segs_size - 1, seg_index);
+ return MIN(vm_page_zones_size - 1, zone_index);
}
static int __init
-vm_page_boot_seg_loaded(const struct vm_page_boot_seg *seg)
+vm_page_boot_zone_loaded(const struct vm_page_boot_zone *zone)
{
- return (seg->end != 0);
+ return (zone->end != 0);
}
static void __init
-vm_page_check_boot_segs(void)
+vm_page_check_boot_zones(void)
{
unsigned int i;
int expect_loaded;
- if (vm_page_segs_size == 0) {
+ if (vm_page_zones_size == 0) {
panic("vm_page: no physical memory loaded");
}
- for (i = 0; i < ARRAY_SIZE(vm_page_boot_segs); i++) {
- expect_loaded = (i < vm_page_segs_size);
+ for (i = 0; i < ARRAY_SIZE(vm_page_boot_zones); i++) {
+ expect_loaded = (i < vm_page_zones_size);
- if (vm_page_boot_seg_loaded(&vm_page_boot_segs[i]) == expect_loaded) {
+ if (vm_page_boot_zone_loaded(&vm_page_boot_zones[i]) == expect_loaded) {
continue;
}
- panic("vm_page: invalid boot segment table");
+ panic("vm_page: invalid boot zone table");
}
}
static phys_addr_t __init
-vm_page_boot_seg_size(struct vm_page_boot_seg *seg)
+vm_page_boot_zone_size(struct vm_page_boot_zone *zone)
{
- return seg->end - seg->start;
+ return zone->end - zone->start;
}
static phys_addr_t __init
-vm_page_boot_seg_avail_size(struct vm_page_boot_seg *seg)
+vm_page_boot_zone_avail_size(struct vm_page_boot_zone *zone)
{
- return seg->avail_end - seg->avail_start;
+ return zone->avail_end - zone->avail_start;
}
static void * __init
vm_page_bootalloc(size_t size)
{
- struct vm_page_boot_seg *seg;
+ struct vm_page_boot_zone *zone;
phys_addr_t pa;
unsigned int i;
- for (i = vm_page_select_alloc_seg(VM_PAGE_SEL_DIRECTMAP);
- i < vm_page_segs_size;
+ for (i = vm_page_select_alloc_zone(VM_PAGE_SEL_DIRECTMAP);
+ i < vm_page_zones_size;
i--) {
- seg = &vm_page_boot_segs[i];
+ zone = &vm_page_boot_zones[i];
- if (!seg->heap_present) {
+ if (!zone->heap_present) {
continue;
}
- if (size <= vm_page_boot_seg_avail_size(seg)) {
- pa = seg->avail_start;
- seg->avail_start += vm_page_round(size);
+ if (size <= vm_page_boot_zone_avail_size(zone)) {
+ pa = zone->avail_start;
+ zone->avail_start += vm_page_round(size);
return (void *)vm_page_direct_va(pa);
}
}
@@ -642,23 +643,23 @@ vm_page_bootalloc(size_t size)
void __init
vm_page_setup(void)
{
- struct vm_page_boot_seg *boot_seg;
- struct vm_page_seg *seg;
+ struct vm_page_boot_zone *boot_zone;
+ struct vm_page_zone *zone;
struct vm_page *table, *page, *end;
size_t nr_pages, table_size;
unsigned long va;
unsigned int i;
phys_addr_t pa;
- vm_page_check_boot_segs();
+ vm_page_check_boot_zones();
/*
* Compute the page table size.
*/
nr_pages = 0;
- for (i = 0; i < vm_page_segs_size; i++) {
- nr_pages += vm_page_atop(vm_page_boot_seg_size(&vm_page_boot_segs[i]));
+ for (i = 0; i < vm_page_zones_size; i++) {
+ nr_pages += vm_page_atop(vm_page_boot_zone_size(&vm_page_boot_zones[i]));
}
table_size = vm_page_round(nr_pages * sizeof(struct vm_page));
@@ -668,26 +669,26 @@ vm_page_setup(void)
va = (unsigned long)table;
/*
- * Initialize the segments, associating them to the page table. When
- * the segments are initialized, all their pages are set allocated.
+ * Initialize the zones, associating them to the page table. When
+ * the zones are initialized, all their pages are set allocated.
* Pages are then released, which populates the free lists.
*/
- for (i = 0; i < vm_page_segs_size; i++) {
- seg = &vm_page_segs[i];
- boot_seg = &vm_page_boot_segs[i];
- vm_page_seg_init(seg, boot_seg->start, boot_seg->end, table);
- page = seg->pages + vm_page_atop(boot_seg->avail_start
- - boot_seg->start);
- end = seg->pages + vm_page_atop(boot_seg->avail_end
- - boot_seg->start);
+ for (i = 0; i < vm_page_zones_size; i++) {
+ zone = &vm_page_zones[i];
+ boot_zone = &vm_page_boot_zones[i];
+ vm_page_zone_init(zone, boot_zone->start, boot_zone->end, table);
+ page = zone->pages + vm_page_atop(boot_zone->avail_start
+ - boot_zone->start);
+ end = zone->pages + vm_page_atop(boot_zone->avail_end
+ - boot_zone->start);
while (page < end) {
page->type = VM_PAGE_FREE;
- vm_page_seg_free_to_buddy(seg, page, 0);
+ vm_page_zone_free_to_buddy(zone, page, 0);
page++;
}
- table += vm_page_atop(vm_page_seg_size(seg));
+ table += vm_page_atop(vm_page_zone_size(zone));
}
while (va < (unsigned long)table) {
@@ -704,24 +705,24 @@ vm_page_setup(void)
void __init
vm_page_manage(struct vm_page *page)
{
- assert(page->seg_index < ARRAY_SIZE(vm_page_segs));
+ assert(page->zone_index < ARRAY_SIZE(vm_page_zones));
assert(page->type == VM_PAGE_RESERVED);
vm_page_set_type(page, 0, VM_PAGE_FREE);
- vm_page_seg_free_to_buddy(&vm_page_segs[page->seg_index], page, 0);
+ vm_page_zone_free_to_buddy(&vm_page_zones[page->zone_index], page, 0);
}
struct vm_page *
vm_page_lookup(phys_addr_t pa)
{
- struct vm_page_seg *seg;
+ struct vm_page_zone *zone;
unsigned int i;
- for (i = 0; i < vm_page_segs_size; i++) {
- seg = &vm_page_segs[i];
+ for (i = 0; i < vm_page_zones_size; i++) {
+ zone = &vm_page_zones[i];
- if ((pa >= seg->start) && (pa < seg->end)) {
- return &seg->pages[vm_page_atop(pa - seg->start)];
+ if ((pa >= zone->start) && (pa < zone->end)) {
+ return &zone->pages[vm_page_atop(pa - zone->start)];
}
}
@@ -734,8 +735,8 @@ vm_page_alloc(unsigned int order, unsigned int selector, unsigned short type)
struct vm_page *page;
unsigned int i;
- for (i = vm_page_select_alloc_seg(selector); i < vm_page_segs_size; i--) {
- page = vm_page_seg_alloc(&vm_page_segs[i], order, type);
+ for (i = vm_page_select_alloc_zone(selector); i < vm_page_zones_size; i--) {
+ page = vm_page_zone_alloc(&vm_page_zones[i], order, type);
if (page != NULL) {
return page;
@@ -748,40 +749,40 @@ vm_page_alloc(unsigned int order, unsigned int selector, unsigned short type)
void
vm_page_free(struct vm_page *page, unsigned int order)
{
- assert(page->seg_index < ARRAY_SIZE(vm_page_segs));
+ assert(page->zone_index < ARRAY_SIZE(vm_page_zones));
- vm_page_seg_free(&vm_page_segs[page->seg_index], page, order);
+ vm_page_zone_free(&vm_page_zones[page->zone_index], page, order);
}
const char *
-vm_page_seg_name(unsigned int seg_index)
+vm_page_zone_name(unsigned int zone_index)
{
- /* Don't use a switch statement since segments can be aliased */
- if (seg_index == VM_PAGE_SEG_HIGHMEM) {
+ /* Don't use a switch statement since zones can be aliased */
+ if (zone_index == VM_PAGE_ZONE_HIGHMEM) {
return "HIGHMEM";
- } else if (seg_index == VM_PAGE_SEG_DIRECTMAP) {
+ } else if (zone_index == VM_PAGE_ZONE_DIRECTMAP) {
return "DIRECTMAP";
- } else if (seg_index == VM_PAGE_SEG_DMA32) {
+ } else if (zone_index == VM_PAGE_ZONE_DMA32) {
return "DMA32";
- } else if (seg_index == VM_PAGE_SEG_DMA) {
+ } else if (zone_index == VM_PAGE_ZONE_DMA) {
return "DMA";
} else {
- panic("vm_page: invalid segment index");
+ panic("vm_page: invalid zone index");
}
}
void
vm_page_info(void)
{
- struct vm_page_seg *seg;
+ struct vm_page_zone *zone;
unsigned long pages;
unsigned int i;
- for (i = 0; i < vm_page_segs_size; i++) {
- seg = &vm_page_segs[i];
- pages = (unsigned long)(seg->pages_end - seg->pages);
+ for (i = 0; i < vm_page_zones_size; i++) {
+ zone = &vm_page_zones[i];
+ pages = (unsigned long)(zone->pages_end - zone->pages);
printk("vm_page: %s: pages: %lu (%luM), free: %lu (%luM)\n",
- vm_page_seg_name(i), pages, pages >> (20 - PAGE_SHIFT),
- seg->nr_free_pages, seg->nr_free_pages >> (20 - PAGE_SHIFT));
+ vm_page_zone_name(i), pages, pages >> (20 - PAGE_SHIFT),
+ zone->nr_free_pages, zone->nr_free_pages >> (20 - PAGE_SHIFT));
}
}