1 /* 2 * Copyright 2002-2009, Axel Dörfler, axeld@pinc-software.de. 3 * Distributed under the terms of the MIT License. 4 * 5 * Copyright 2001-2002, Travis Geiselbrecht. All rights reserved. 6 * Distributed under the terms of the NewOS License. 7 */ 8 #ifndef _KERNEL_VM_VM_H 9 #define _KERNEL_VM_VM_H 10 11 #include <OS.h> 12 13 #include <arch/vm.h> 14 #include <vm_defs.h> 15 #include <vm/vm_types.h> 16 17 18 struct generic_io_vec; 19 struct kernel_args; 20 struct ObjectCache; 21 struct system_memory_info; 22 struct team; 23 struct VMAddressSpace; 24 struct VMArea; 25 struct VMCache; 26 struct vm_page; 27 struct vnode; 28 struct VMPageWiringInfo; 29 30 31 // area creation flags 32 #define CREATE_AREA_DONT_WAIT 0x01 33 #define CREATE_AREA_UNMAP_ADDRESS_RANGE 0x02 34 #define CREATE_AREA_DONT_CLEAR 0x04 35 #define CREATE_AREA_PRIORITY_VIP 0x08 36 #define CREATE_AREA_DONT_COMMIT_MEMORY 0x10 37 38 // memory/page allocation priorities 39 #define VM_PRIORITY_USER 0 40 #define VM_PRIORITY_SYSTEM 1 41 #define VM_PRIORITY_VIP 2 42 43 // page reserves 44 #define VM_PAGE_RESERVE_USER 512 45 #define VM_PAGE_RESERVE_SYSTEM 128 46 47 // memory reserves 48 #define VM_MEMORY_RESERVE_USER (VM_PAGE_RESERVE_USER * B_PAGE_SIZE) 49 #define VM_MEMORY_RESERVE_SYSTEM (VM_PAGE_RESERVE_SYSTEM * B_PAGE_SIZE) 50 51 52 extern struct ObjectCache* gPageMappingsObjectCache; 53 54 55 #ifdef __cplusplus 56 extern "C" { 57 #endif 58 59 // startup only 60 status_t vm_init(struct kernel_args *args); 61 status_t vm_init_post_sem(struct kernel_args *args); 62 status_t vm_init_post_thread(struct kernel_args *args); 63 status_t vm_init_post_modules(struct kernel_args *args); 64 void vm_free_kernel_args(struct kernel_args *args); 65 void vm_free_unused_boot_loader_range(addr_t start, addr_t end); 66 page_num_t vm_allocate_early_physical_page(kernel_args *args); 67 addr_t vm_allocate_early(struct kernel_args *args, size_t virtualSize, 68 size_t physicalSize, uint32 attributes, addr_t alignment); 69 70 void slab_init(struct kernel_args *args); 71 void slab_init_post_area(); 72 void slab_init_post_sem(); 73 void slab_init_post_thread(); 74 75 // to protect code regions with interrupts turned on 76 void permit_page_faults(void); 77 void forbid_page_faults(void); 78 79 // private kernel only extension (should be moved somewhere else): 80 area_id create_area_etc(team_id team, const char *name, uint32 size, 81 uint32 lock, uint32 protection, uint32 flags, 82 const virtual_address_restrictions* virtualAddressRestrictions, 83 const physical_address_restrictions* physicalAddressRestrictions, 84 void **_address); 85 area_id transfer_area(area_id id, void** _address, uint32 addressSpec, 86 team_id target, bool kernel); 87 88 const char* vm_cache_type_to_string(int32 type); 89 90 status_t vm_block_address_range(const char* name, void* address, addr_t size); 91 status_t vm_unreserve_address_range(team_id team, void *address, addr_t size); 92 status_t vm_reserve_address_range(team_id team, void **_address, 93 uint32 addressSpec, addr_t size, uint32 flags); 94 area_id vm_create_anonymous_area(team_id team, const char* name, addr_t size, 95 uint32 wiring, uint32 protection, uint32 flags, 96 const virtual_address_restrictions* virtualAddressRestrictions, 97 const physical_address_restrictions* physicalAddressRestrictions, 98 bool kernel, void** _address); 99 area_id vm_map_physical_memory(team_id team, const char *name, void **address, 100 uint32 addressSpec, addr_t size, uint32 protection, 101 phys_addr_t physicalAddress, bool alreadyWired); 102 area_id vm_map_physical_memory_vecs(team_id team, const char* name, 103 void** _address, uint32 addressSpec, addr_t* _size, uint32 protection, 104 struct generic_io_vec* vecs, uint32 vecCount); 105 area_id vm_map_file(team_id aid, const char *name, void **address, 106 uint32 addressSpec, addr_t size, uint32 protection, uint32 mapping, 107 bool unmapAddressRange, int fd, off_t offset); 108 struct VMCache *vm_area_get_locked_cache(struct VMArea *area); 109 void vm_area_put_locked_cache(struct VMCache *cache); 110 area_id vm_create_null_area(team_id team, const char *name, void **address, 111 uint32 addressSpec, addr_t size, uint32 flags); 112 area_id vm_copy_area(team_id team, const char *name, void **_address, 113 uint32 addressSpec, uint32 protection, area_id sourceID); 114 area_id vm_clone_area(team_id team, const char *name, void **address, 115 uint32 addressSpec, uint32 protection, uint32 mapping, 116 area_id sourceArea, bool kernel); 117 status_t vm_delete_area(team_id teamID, area_id areaID, bool kernel); 118 status_t vm_create_vnode_cache(struct vnode *vnode, struct VMCache **_cache); 119 status_t vm_set_area_memory_type(area_id id, phys_addr_t physicalBase, 120 uint32 type); 121 status_t vm_get_page_mapping(team_id team, addr_t vaddr, phys_addr_t *paddr); 122 bool vm_test_map_modification(struct vm_page *page); 123 void vm_clear_map_flags(struct vm_page *page, uint32 flags); 124 void vm_remove_all_page_mappings(struct vm_page *page); 125 int32 vm_clear_page_mapping_accessed_flags(struct vm_page *page); 126 int32 vm_remove_all_page_mappings_if_unaccessed(struct vm_page *page); 127 status_t vm_wire_page(team_id team, addr_t address, bool writable, 128 struct VMPageWiringInfo* info); 129 void vm_unwire_page(struct VMPageWiringInfo* info); 130 131 status_t vm_get_physical_page(phys_addr_t paddr, addr_t* vaddr, void** _handle); 132 status_t vm_put_physical_page(addr_t vaddr, void* handle); 133 status_t vm_get_physical_page_current_cpu(phys_addr_t paddr, addr_t* vaddr, 134 void** _handle); 135 status_t vm_put_physical_page_current_cpu(addr_t vaddr, void* handle); 136 status_t vm_get_physical_page_debug(phys_addr_t paddr, addr_t* vaddr, 137 void** _handle); 138 status_t vm_put_physical_page_debug(addr_t vaddr, void* handle); 139 140 void vm_get_info(struct system_memory_info *info); 141 uint32 vm_num_page_faults(void); 142 off_t vm_available_memory(void); 143 off_t vm_available_not_needed_memory(void); 144 off_t vm_available_not_needed_memory_debug(void); 145 size_t vm_kernel_address_space_left(void); 146 147 status_t vm_memset_physical(phys_addr_t address, int value, size_t length); 148 status_t vm_memcpy_from_physical(void* to, phys_addr_t from, size_t length, 149 bool user); 150 status_t vm_memcpy_to_physical(phys_addr_t to, const void* from, size_t length, 151 bool user); 152 void vm_memcpy_physical_page(phys_addr_t to, phys_addr_t from); 153 154 status_t vm_debug_copy_page_memory(team_id teamID, void* unsafeMemory, 155 void* buffer, size_t size, bool copyToUnsafe); 156 157 // user syscalls 158 area_id _user_create_area(const char *name, void **address, uint32 addressSpec, 159 size_t size, uint32 lock, uint32 protection); 160 status_t _user_delete_area(area_id area); 161 162 area_id _user_map_file(const char *uname, void **uaddress, uint32 addressSpec, 163 size_t size, uint32 protection, uint32 mapping, 164 bool unmapAddressRange, int fd, off_t offset); 165 status_t _user_unmap_memory(void *address, size_t size); 166 status_t _user_set_memory_protection(void* address, size_t size, 167 uint32 protection); 168 status_t _user_sync_memory(void *address, size_t size, uint32 flags); 169 status_t _user_memory_advice(void* address, size_t size, uint32 advice); 170 171 area_id _user_area_for(void *address); 172 area_id _user_find_area(const char *name); 173 status_t _user_get_area_info(area_id area, area_info *info); 174 status_t _user_get_next_area_info(team_id team, int32 *cookie, area_info *info); 175 status_t _user_resize_area(area_id area, size_t newSize); 176 area_id _user_transfer_area(area_id area, void **_address, uint32 addressSpec, 177 team_id target); 178 status_t _user_set_area_protection(area_id area, uint32 newProtection); 179 area_id _user_clone_area(const char *name, void **_address, uint32 addressSpec, 180 uint32 protection, area_id sourceArea); 181 status_t _user_reserve_address_range(addr_t* userAddress, uint32 addressSpec, 182 addr_t size); 183 status_t _user_unreserve_address_range(addr_t address, addr_t size); 184 185 #ifdef __cplusplus 186 } 187 #endif 188 189 #endif /* _KERNEL_VM_VM_H */ 190