summaryrefslogtreecommitdiff
path: root/vm
diff options
context:
space:
mode:
authorRichard Braun <rbraun@sceen.net>2012-10-10 21:18:28 +0200
committerRichard Braun <rbraun@sceen.net>2012-10-10 21:18:28 +0200
commitd754647055848c8d575e96d09cdba9e16b8f27ba (patch)
treedc5094de4fc94df137cda5d745fedbd15319e51d /vm
parent8eadbf1e9f3530fa7b0f93caf572d96a50891445 (diff)
Rename vm_phys_t to phys_addr_t
Diffstat (limited to 'vm')
-rw-r--r--vm/vm_kmem.c8
-rw-r--r--vm/vm_kmem.h2
-rw-r--r--vm/vm_page.h4
-rw-r--r--vm/vm_phys.c28
-rw-r--r--vm/vm_phys.h8
5 files changed, 25 insertions, 25 deletions
diff --git a/vm/vm_kmem.c b/vm/vm_kmem.c
index c9fd4027..8617791d 100644
--- a/vm/vm_kmem.c
+++ b/vm/vm_kmem.c
@@ -51,7 +51,7 @@ unsigned long __init
vm_kmem_bootalloc(size_t size)
{
unsigned long start, va;
- vm_phys_t pa;
+ phys_addr_t pa;
assert(size > 0);
@@ -162,7 +162,7 @@ vm_kmem_free(unsigned long addr, size_t size)
{
struct vm_page *page;
unsigned long va, end;
- vm_phys_t pa;
+ phys_addr_t pa;
assert(vm_kmem_free_check(addr, size) == 0);
@@ -184,12 +184,12 @@ vm_kmem_free(unsigned long addr, size_t size)
}
void *
-vm_kmem_map_pa(vm_phys_t addr, size_t size, unsigned long *map_addrp,
+vm_kmem_map_pa(phys_addr_t addr, size_t size, unsigned long *map_addrp,
size_t *map_sizep)
{
unsigned long offset, map_addr;
size_t map_size;
- vm_phys_t start;
+ phys_addr_t start;
assert(vm_kmem_alloc_check(size) == 0);
diff --git a/vm/vm_kmem.h b/vm/vm_kmem.h
index d4984e8e..6ecf8e17 100644
--- a/vm/vm_kmem.h
+++ b/vm/vm_kmem.h
@@ -74,7 +74,7 @@ void vm_kmem_free(unsigned long addr, size_t size);
* TODO When mapping attributes are implemented, make this function disable
* caching on the mapping.
*/
-void * vm_kmem_map_pa(vm_phys_t addr, size_t size, unsigned long *map_addrp,
+void * vm_kmem_map_pa(phys_addr_t addr, size_t size, unsigned long *map_addrp,
size_t *map_sizep);
/*
diff --git a/vm/vm_page.h b/vm/vm_page.h
index 641cbccf..e499dc52 100644
--- a/vm/vm_page.h
+++ b/vm/vm_page.h
@@ -41,10 +41,10 @@ struct vm_page {
struct list node;
unsigned short seg_index;
unsigned short order;
- vm_phys_t phys_addr;
+ phys_addr_t phys_addr;
};
-static inline vm_phys_t
+static inline phys_addr_t
vm_page_to_pa(const struct vm_page *page)
{
return page->phys_addr;
diff --git a/vm/vm_phys.c b/vm/vm_phys.c
index 2852a034..81e655bf 100644
--- a/vm/vm_phys.c
+++ b/vm/vm_phys.c
@@ -105,8 +105,8 @@ struct vm_phys_seg {
struct vm_phys_cpu_pool cpu_pools[MAX_CPUS];
struct list node;
- vm_phys_t start;
- vm_phys_t end;
+ phys_addr_t start;
+ phys_addr_t end;
struct vm_page *pages;
struct vm_page *pages_end;
/* struct mutex mutex; */
@@ -119,8 +119,8 @@ struct vm_phys_seg {
* Bootstrap information about a segment.
*/
struct vm_phys_boot_seg {
- vm_phys_t avail_start;
- vm_phys_t avail_end;
+ phys_addr_t avail_start;
+ phys_addr_t avail_end;
};
int vm_phys_ready;
@@ -150,7 +150,7 @@ static int vm_phys_load_initialized __initdata = 0;
static void __init
vm_phys_init_page(struct vm_page *page, unsigned short seg_index,
- unsigned short order, vm_phys_t pa)
+ unsigned short order, phys_addr_t pa)
{
page->seg_index = seg_index;
page->order = order;
@@ -225,7 +225,7 @@ vm_phys_seg_free_to_buddy(struct vm_phys_seg *seg, struct vm_page *page,
unsigned int order)
{
struct vm_page *buddy;
- vm_phys_t pa, buddy_pa;
+ phys_addr_t pa, buddy_pa;
unsigned int nr_pages;
assert(page >= seg->pages);
@@ -339,7 +339,7 @@ vm_phys_cpu_pool_drain(struct vm_phys_cpu_pool *cpu_pool,
/* mutex_unlock(&seg->mutex); */
}
-static inline vm_phys_t __init
+static inline phys_addr_t __init
vm_phys_seg_size(struct vm_phys_seg *seg)
{
return seg->end - seg->start;
@@ -348,7 +348,7 @@ vm_phys_seg_size(struct vm_phys_seg *seg)
static int __init
vm_phys_seg_compute_pool_size(struct vm_phys_seg *seg)
{
- vm_phys_t size;
+ phys_addr_t size;
size = vm_page_atop(vm_phys_seg_size(seg)) / VM_PHYS_CPU_POOL_RATIO;
@@ -363,7 +363,7 @@ vm_phys_seg_compute_pool_size(struct vm_phys_seg *seg)
static void __init
vm_phys_seg_init(struct vm_phys_seg *seg, struct vm_page *pages)
{
- vm_phys_t pa;
+ phys_addr_t pa;
int pool_size;
unsigned int i;
@@ -447,8 +447,8 @@ vm_phys_seg_free(struct vm_phys_seg *seg, struct vm_page *page,
}
void __init
-vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end,
- vm_phys_t avail_start, vm_phys_t avail_end,
+vm_phys_load(const char *name, phys_addr_t start, phys_addr_t end,
+ phys_addr_t avail_start, phys_addr_t avail_end,
unsigned int seglist_prio)
{
struct vm_phys_boot_seg *boot_seg;
@@ -485,13 +485,13 @@ vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end,
vm_phys_segs_size++;
}
-vm_phys_t __init
+phys_addr_t __init
vm_phys_bootalloc(void)
{
struct vm_phys_boot_seg *boot_seg;
struct vm_phys_seg *seg;
struct list *seg_list;
- vm_phys_t pa;
+ phys_addr_t pa;
for (seg_list = &vm_phys_seg_lists[ARRAY_SIZE(vm_phys_seg_lists) - 1];
seg_list >= vm_phys_seg_lists;
@@ -565,7 +565,7 @@ vm_phys_manage(struct vm_page *page)
}
struct vm_page *
-vm_phys_lookup_page(vm_phys_t pa)
+vm_phys_lookup_page(phys_addr_t pa)
{
struct vm_phys_seg *seg;
unsigned int i;
diff --git a/vm/vm_phys.h b/vm/vm_phys.h
index a5a7d32f..aff124a0 100644
--- a/vm/vm_phys.h
+++ b/vm/vm_phys.h
@@ -36,8 +36,8 @@ extern int vm_phys_ready;
* The avail_start and avail_end parameters are used to maintain a simple
* heap for bootstrap allocations.
*/
-void vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end,
- vm_phys_t avail_start, vm_phys_t avail_end,
+void vm_phys_load(const char *name, phys_addr_t start, phys_addr_t end,
+ phys_addr_t avail_start, phys_addr_t avail_end,
unsigned int seglist_prio);
/*
@@ -46,7 +46,7 @@ void vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end,
* This function is used to allocate physical memory at boot time, before the
* vm_phys module is ready, but after the physical memory has been loaded.
*/
-vm_phys_t vm_phys_bootalloc(void);
+phys_addr_t vm_phys_bootalloc(void);
/*
* Set up the vm_phys module.
@@ -67,7 +67,7 @@ void vm_phys_manage(struct vm_page *page);
/*
* Return the page descriptor for the given physical address.
*/
-struct vm_page * vm_phys_lookup_page(vm_phys_t pa);
+struct vm_page * vm_phys_lookup_page(phys_addr_t pa);
/*
* Allocate a block of 2^order physical pages.