diff options
author | Richard Braun <rbraun@sceen.net> | 2012-10-10 21:18:28 +0200 |
---|---|---|
committer | Richard Braun <rbraun@sceen.net> | 2012-10-10 21:18:28 +0200 |
commit | d754647055848c8d575e96d09cdba9e16b8f27ba (patch) | |
tree | dc5094de4fc94df137cda5d745fedbd15319e51d /vm | |
parent | 8eadbf1e9f3530fa7b0f93caf572d96a50891445 (diff) |
Rename vm_phys_t to phys_addr_t
Diffstat (limited to 'vm')
-rw-r--r-- | vm/vm_kmem.c | 8 | ||||
-rw-r--r-- | vm/vm_kmem.h | 2 | ||||
-rw-r--r-- | vm/vm_page.h | 4 | ||||
-rw-r--r-- | vm/vm_phys.c | 28 | ||||
-rw-r--r-- | vm/vm_phys.h | 8 |
5 files changed, 25 insertions, 25 deletions
diff --git a/vm/vm_kmem.c b/vm/vm_kmem.c index c9fd4027..8617791d 100644 --- a/vm/vm_kmem.c +++ b/vm/vm_kmem.c @@ -51,7 +51,7 @@ unsigned long __init vm_kmem_bootalloc(size_t size) { unsigned long start, va; - vm_phys_t pa; + phys_addr_t pa; assert(size > 0); @@ -162,7 +162,7 @@ vm_kmem_free(unsigned long addr, size_t size) { struct vm_page *page; unsigned long va, end; - vm_phys_t pa; + phys_addr_t pa; assert(vm_kmem_free_check(addr, size) == 0); @@ -184,12 +184,12 @@ vm_kmem_free(unsigned long addr, size_t size) } void * -vm_kmem_map_pa(vm_phys_t addr, size_t size, unsigned long *map_addrp, +vm_kmem_map_pa(phys_addr_t addr, size_t size, unsigned long *map_addrp, size_t *map_sizep) { unsigned long offset, map_addr; size_t map_size; - vm_phys_t start; + phys_addr_t start; assert(vm_kmem_alloc_check(size) == 0); diff --git a/vm/vm_kmem.h b/vm/vm_kmem.h index d4984e8e..6ecf8e17 100644 --- a/vm/vm_kmem.h +++ b/vm/vm_kmem.h @@ -74,7 +74,7 @@ void vm_kmem_free(unsigned long addr, size_t size); * TODO When mapping attributes are implemented, make this function disable * caching on the mapping. */ -void * vm_kmem_map_pa(vm_phys_t addr, size_t size, unsigned long *map_addrp, +void * vm_kmem_map_pa(phys_addr_t addr, size_t size, unsigned long *map_addrp, size_t *map_sizep); /* diff --git a/vm/vm_page.h b/vm/vm_page.h index 641cbccf..e499dc52 100644 --- a/vm/vm_page.h +++ b/vm/vm_page.h @@ -41,10 +41,10 @@ struct vm_page { struct list node; unsigned short seg_index; unsigned short order; - vm_phys_t phys_addr; + phys_addr_t phys_addr; }; -static inline vm_phys_t +static inline phys_addr_t vm_page_to_pa(const struct vm_page *page) { return page->phys_addr; diff --git a/vm/vm_phys.c b/vm/vm_phys.c index 2852a034..81e655bf 100644 --- a/vm/vm_phys.c +++ b/vm/vm_phys.c @@ -105,8 +105,8 @@ struct vm_phys_seg { struct vm_phys_cpu_pool cpu_pools[MAX_CPUS]; struct list node; - vm_phys_t start; - vm_phys_t end; + phys_addr_t start; + phys_addr_t end; struct vm_page *pages; struct vm_page *pages_end; /* struct mutex mutex; */ @@ -119,8 +119,8 @@ struct vm_phys_seg { * Bootstrap information about a segment. */ struct vm_phys_boot_seg { - vm_phys_t avail_start; - vm_phys_t avail_end; + phys_addr_t avail_start; + phys_addr_t avail_end; }; int vm_phys_ready; @@ -150,7 +150,7 @@ static int vm_phys_load_initialized __initdata = 0; static void __init vm_phys_init_page(struct vm_page *page, unsigned short seg_index, - unsigned short order, vm_phys_t pa) + unsigned short order, phys_addr_t pa) { page->seg_index = seg_index; page->order = order; @@ -225,7 +225,7 @@ vm_phys_seg_free_to_buddy(struct vm_phys_seg *seg, struct vm_page *page, unsigned int order) { struct vm_page *buddy; - vm_phys_t pa, buddy_pa; + phys_addr_t pa, buddy_pa; unsigned int nr_pages; assert(page >= seg->pages); @@ -339,7 +339,7 @@ vm_phys_cpu_pool_drain(struct vm_phys_cpu_pool *cpu_pool, /* mutex_unlock(&seg->mutex); */ } -static inline vm_phys_t __init +static inline phys_addr_t __init vm_phys_seg_size(struct vm_phys_seg *seg) { return seg->end - seg->start; @@ -348,7 +348,7 @@ vm_phys_seg_size(struct vm_phys_seg *seg) static int __init vm_phys_seg_compute_pool_size(struct vm_phys_seg *seg) { - vm_phys_t size; + phys_addr_t size; size = vm_page_atop(vm_phys_seg_size(seg)) / VM_PHYS_CPU_POOL_RATIO; @@ -363,7 +363,7 @@ vm_phys_seg_compute_pool_size(struct vm_phys_seg *seg) static void __init vm_phys_seg_init(struct vm_phys_seg *seg, struct vm_page *pages) { - vm_phys_t pa; + phys_addr_t pa; int pool_size; unsigned int i; @@ -447,8 +447,8 @@ vm_phys_seg_free(struct vm_phys_seg *seg, struct vm_page *page, } void __init -vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end, - vm_phys_t avail_start, vm_phys_t avail_end, +vm_phys_load(const char *name, phys_addr_t start, phys_addr_t end, + phys_addr_t avail_start, phys_addr_t avail_end, unsigned int seglist_prio) { struct vm_phys_boot_seg *boot_seg; @@ -485,13 +485,13 @@ vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end, vm_phys_segs_size++; } -vm_phys_t __init +phys_addr_t __init vm_phys_bootalloc(void) { struct vm_phys_boot_seg *boot_seg; struct vm_phys_seg *seg; struct list *seg_list; - vm_phys_t pa; + phys_addr_t pa; for (seg_list = &vm_phys_seg_lists[ARRAY_SIZE(vm_phys_seg_lists) - 1]; seg_list >= vm_phys_seg_lists; @@ -565,7 +565,7 @@ vm_phys_manage(struct vm_page *page) } struct vm_page * -vm_phys_lookup_page(vm_phys_t pa) +vm_phys_lookup_page(phys_addr_t pa) { struct vm_phys_seg *seg; unsigned int i; diff --git a/vm/vm_phys.h b/vm/vm_phys.h index a5a7d32f..aff124a0 100644 --- a/vm/vm_phys.h +++ b/vm/vm_phys.h @@ -36,8 +36,8 @@ extern int vm_phys_ready; * The avail_start and avail_end parameters are used to maintain a simple * heap for bootstrap allocations. */ -void vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end, - vm_phys_t avail_start, vm_phys_t avail_end, +void vm_phys_load(const char *name, phys_addr_t start, phys_addr_t end, + phys_addr_t avail_start, phys_addr_t avail_end, unsigned int seglist_prio); /* @@ -46,7 +46,7 @@ void vm_phys_load(const char *name, vm_phys_t start, vm_phys_t end, * This function is used to allocate physical memory at boot time, before the * vm_phys module is ready, but after the physical memory has been loaded. */ -vm_phys_t vm_phys_bootalloc(void); +phys_addr_t vm_phys_bootalloc(void); /* * Set up the vm_phys module. @@ -67,7 +67,7 @@ void vm_phys_manage(struct vm_page *page); /* * Return the page descriptor for the given physical address. */ -struct vm_page * vm_phys_lookup_page(vm_phys_t pa); +struct vm_page * vm_phys_lookup_page(phys_addr_t pa); /* * Allocate a block of 2^order physical pages. |