/*
* Copyright (c) 2010-2017 Richard Braun.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see .
*
*
* Physical page management.
*
* A page is said to be managed if it's linked to a VM object, in which
* case there is at least one reference to it.
*/
#ifndef VM_VM_PAGE_H
#define VM_VM_PAGE_H
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
/*
* Byte/page conversion and rounding macros (not inline functions to
* be easily usable on both virtual and physical addresses, which may not
* have the same type size).
*/
#define vm_page_btop(bytes) ((bytes) >> PAGE_SHIFT)
#define vm_page_ptob(pages) ((pages) << PAGE_SHIFT)
#define vm_page_trunc(bytes) P2ALIGN(bytes, PAGE_SIZE)
#define vm_page_round(bytes) P2ROUND(bytes, PAGE_SIZE)
#define vm_page_end(bytes) P2END(bytes, PAGE_SIZE)
#define vm_page_aligned(bytes) P2ALIGNED(bytes, PAGE_SIZE)
/*
* Zone selectors.
*
* Selector-to-zone-list translation table :
* DMA DMA
* DMA32 DMA32 DMA
* DIRECTMAP DIRECTMAP DMA32 DMA
* HIGHMEM HIGHMEM DIRECTMAP DMA32 DMA
*/
#define VM_PAGE_SEL_DMA 0
#define VM_PAGE_SEL_DMA32 1
#define VM_PAGE_SEL_DIRECTMAP 2
#define VM_PAGE_SEL_HIGHMEM 3
/*
* Page usage types.
*/
#define VM_PAGE_FREE 0 /* Page unused */
#define VM_PAGE_RESERVED 1 /* Page reserved at boot time */
#define VM_PAGE_TABLE 2 /* Page is part of the page table */
#define VM_PAGE_PMAP 3 /* Page stores pmap-specific data */
#define VM_PAGE_KMEM 4 /* Page is a direct-mapped kmem slab */
#define VM_PAGE_OBJECT 5 /* Page is part of a VM object */
#define VM_PAGE_KERNEL 6 /* Type for generic kernel allocations */
/*
* Physical page descriptor.
*/
struct vm_page {
struct list node;
unsigned short type;
unsigned short zone_index;
unsigned short order;
phys_addr_t phys_addr;
void *priv;
unsigned int nr_refs;
/* VM object back reference */
struct vm_object *object;
uint64_t offset;
};
static inline unsigned short
vm_page_type(const struct vm_page *page)
{
return page->type;
}
void vm_page_set_type(struct vm_page *page, unsigned int order,
unsigned short type);
static inline unsigned int
vm_page_order(size_t size)
{
return log2_order(vm_page_btop(vm_page_round(size)));
}
static inline phys_addr_t
vm_page_to_pa(const struct vm_page *page)
{
return page->phys_addr;
}
static inline uintptr_t
vm_page_direct_va(phys_addr_t pa)
{
assert(pa < PMEM_DIRECTMAP_LIMIT);
return ((uintptr_t)pa + PMAP_START_DIRECTMAP_ADDRESS);
}
static inline phys_addr_t
vm_page_direct_pa(uintptr_t va)
{
assert(va >= PMAP_START_DIRECTMAP_ADDRESS);
assert(va < PMAP_END_DIRECTMAP_ADDRESS);
return (va - PMAP_START_DIRECTMAP_ADDRESS);
}
static inline void *
vm_page_direct_ptr(const struct vm_page *page)
{
return (void *)vm_page_direct_va(vm_page_to_pa(page));
}
/*
* Associate private data with a page.
*/
static inline void
vm_page_set_priv(struct vm_page *page, void *priv)
{
page->priv = priv;
}
static inline void *
vm_page_get_priv(const struct vm_page *page)
{
return page->priv;
}
static inline void
vm_page_link(struct vm_page *page, struct vm_object *object, uint64_t offset)
{
assert(object != NULL);
page->object = object;
page->offset = offset;
}
static inline void
vm_page_unlink(struct vm_page *page)
{
assert(page->object != NULL);
page->object = NULL;
}
/*
* Load physical memory into the vm_page module at boot time.
*
* All addresses must be page-aligned. Zones can be loaded in any order.
*/
void vm_page_load(unsigned int zone_index, phys_addr_t start, phys_addr_t end);
/*
* Load available physical memory into the vm_page module at boot time.
*
* The zone referred to must have been loaded with vm_page_load
* before loading its heap.
*/
void vm_page_load_heap(unsigned int zone_index, phys_addr_t start,
phys_addr_t end);
/*
* Return true if the vm_page module is completely initialized, false
* otherwise, in which case only vm_page_bootalloc() can be used for
* allocations.
*/
int vm_page_ready(void);
/*
* Make the given page managed by the vm_page module.
*
* If additional memory can be made usable after the VM system is initialized,
* it should be reported through this function.
*/
void vm_page_manage(struct vm_page *page);
/*
* Return the page descriptor for the given physical address.
*/
struct vm_page * vm_page_lookup(phys_addr_t pa);
/*
* Allocate a block of 2^order physical pages.
*
* The selector is used to determine the zones from which allocation can
* be attempted.
*
* If successful, the returned pages have no references.
*/
struct vm_page * vm_page_alloc(unsigned int order, unsigned int selector,
unsigned short type);
/*
* Release a block of 2^order physical pages.
*
* The pages must have no references.
*/
void vm_page_free(struct vm_page *page, unsigned int order);
/*
* Return the name of the given zone.
*/
const char * vm_page_zone_name(unsigned int zone_index);
/*
* Log information about physical pages.
*/
void vm_page_log_info(void);
static inline bool
vm_page_referenced(const struct vm_page *page)
{
return atomic_load(&page->nr_refs, ATOMIC_RELAXED) != 0;
}
static inline void
vm_page_ref(struct vm_page *page)
{
unsigned int nr_refs;
nr_refs = atomic_fetch_add(&page->nr_refs, 1, ATOMIC_RELAXED);
assert(nr_refs != (unsigned int)-1);
}
static inline void
vm_page_unref(struct vm_page *page)
{
unsigned int nr_refs;
nr_refs = atomic_fetch_sub(&page->nr_refs, 1, ATOMIC_ACQ_REL);
assert(nr_refs != 0);
if (nr_refs == 1) {
vm_page_free(page, 0);
}
}
static inline int
vm_page_tryref(struct vm_page *page)
{
unsigned int nr_refs, prev;
do {
nr_refs = atomic_load(&page->nr_refs, ATOMIC_RELAXED);
if (nr_refs == 0) {
return EAGAIN;
}
prev = atomic_cas(&page->nr_refs, nr_refs,
nr_refs + 1, ATOMIC_ACQUIRE);
} while (prev != nr_refs);
return 0;
}
/*
* This init operation provides :
* - module fully initialized
*/
INIT_OP_DECLARE(vm_page_setup);
#endif /* VM_VM_PAGE_H */