50e2064049
This commit removes all traces of Minix segments (the text/data/stack memory map abstraction in the kernel) and significance of Intel segments (hardware segments like CS, DS that add offsets to all addressing before page table translation). This ultimately simplifies the memory layout and addressing and makes the same layout possible on non-Intel architectures. There are only two types of addresses in the world now: virtual and physical; even the kernel and processes have the same virtual address space. Kernel and user processes can be distinguished at a glance as processes won't use 0xF0000000 and above. No static pre-allocated memory sizes exist any more. Changes to booting: . The pre_init.c leaves the kernel and modules exactly as they were left by the bootloader in physical memory . The kernel starts running using physical addressing, loaded at a fixed location given in its linker script by the bootloader. All code and data in this phase are linked to this fixed low location. . It makes a bootstrap pagetable to map itself to a fixed high location (also in linker script) and jumps to the high address. All code and data then use this high addressing. . All code/data symbols linked at the low addresses is prefixed by an objcopy step with __k_unpaged_*, so that that code cannot reference highly-linked symbols (which aren't valid yet) or vice versa (symbols that aren't valid any more). . The two addressing modes are separated in the linker script by collecting the unpaged_*.o objects and linking them with low addresses, and linking the rest high. Some objects are linked twice, once low and once high. . The bootstrap phase passes a lot of information (e.g. free memory list, physical location of the modules, etc.) using the kinfo struct. . After this bootstrap the low-linked part is freed. . The kernel maps in VM into the bootstrap page table so that VM can begin executing. Its first job is to make page tables for all other boot processes. So VM runs before RS, and RS gets a fully dynamic, VM-managed address space. VM gets its privilege info from RS as usual but that happens after RS starts running. . Both the kernel loading VM and VM organizing boot processes happen using the libexec logic. This removes the last reason for VM to still know much about exec() and vm/exec.c is gone. Further Implementation: . All segments are based at 0 and have a 4 GB limit. . The kernel is mapped in at the top of the virtual address space so as not to constrain the user processes. . Processes do not use segments from the LDT at all; there are no segments in the LDT any more, so no LLDT is needed. . The Minix segments T/D/S are gone and so none of the user-space or in-kernel copy functions use them. The copy functions use a process endpoint of NONE to realize it's a physical address, virtual otherwise. . The umap call only makes sense to translate a virtual address to a physical address now. . Segments-related calls like newmap and alloc_segments are gone. . All segments-related translation in VM is gone (vir2map etc). . Initialization in VM is simpler as no moving around is necessary. . VM and all other boot processes can be linked wherever they wish and will be mapped in at the right location by the kernel and VM respectively. Other changes: . The multiboot code is less special: it does not use mb_print for its diagnostics any more but uses printf() as normal, saving the output into the diagnostics buffer, only printing to the screen using the direct print functions if a panic() occurs. . The multiboot code uses the flexible 'free memory map list' style to receive the list of free memory if available. . The kernel determines the memory layout of the processes to a degree: it tells VM where the kernel starts and ends and where the kernel wants the top of the process to be. VM then uses this entire range, i.e. the stack is right at the top, and mmap()ped bits of memory are placed below that downwards, and the break grows upwards. Other Consequences: . Every process gets its own page table as address spaces can't be separated any more by segments. . As all segments are 0-based, there is no distinction between virtual and linear addresses, nor between userspace and kernel addresses. . Less work is done when context switching, leading to a net performance increase. (8% faster on my machine for 'make servers'.) . The layout and configuration of the GDT makes sysenter and syscall possible.
194 lines
6.4 KiB
C
194 lines
6.4 KiB
C
/* Function prototypes. */
|
|
|
|
struct vmproc;
|
|
struct stat;
|
|
struct memory;
|
|
struct vir_region;
|
|
struct phys_region;
|
|
|
|
#include <minix/ipc.h>
|
|
#include <minix/endpoint.h>
|
|
#include <minix/safecopies.h>
|
|
#include <minix/vm.h>
|
|
#include <timers.h>
|
|
#include <stdio.h>
|
|
#include <pagetable.h>
|
|
|
|
#include "vm.h"
|
|
#include "yielded.h"
|
|
|
|
/* alloc.c */
|
|
void mem_sanitycheck(char *file, int line);
|
|
phys_clicks alloc_mem(phys_clicks clicks, u32_t flags);
|
|
struct memlist *alloc_mem_in_list(phys_bytes bytes, u32_t flags);
|
|
int do_adddma(message *msg);
|
|
int do_deldma(message *msg);
|
|
int do_getdma(message *msg);
|
|
void release_dma(struct vmproc *vmp);
|
|
void memstats(int *nodes, int *pages, int *largest);
|
|
void printmemstats(void);
|
|
void usedpages_reset(void);
|
|
int usedpages_add_f(phys_bytes phys, phys_bytes len, char *file, int
|
|
line);
|
|
void free_mem(phys_clicks base, phys_clicks clicks);
|
|
void free_mem_list(struct memlist *list, int all);
|
|
void print_mem_list(struct memlist *ml);
|
|
#define usedpages_add(a, l) usedpages_add_f(a, l, __FILE__, __LINE__)
|
|
|
|
void mem_init(struct memory *chunks);
|
|
|
|
/* utility.c */
|
|
void get_mem_chunks(struct memory *mem_chunks);
|
|
void reserve_proc_mem(struct memory *mem_chunks, struct mem_map
|
|
*map_ptr);
|
|
int vm_isokendpt(endpoint_t ep, int *proc);
|
|
int get_stack_ptr(int proc_nr, vir_bytes *sp);
|
|
int do_info(message *);
|
|
int swap_proc_slot(struct vmproc *src_vmp, struct vmproc *dst_vmp);
|
|
int swap_proc_dyn_data(struct vmproc *src_vmp, struct vmproc *dst_vmp);
|
|
|
|
/* exit.c */
|
|
void clear_proc(struct vmproc *vmp);
|
|
int do_exit(message *msg);
|
|
int do_willexit(message *msg);
|
|
int do_procctl(message *msg);
|
|
void free_proc(struct vmproc *vmp);
|
|
|
|
/* fork.c */
|
|
int do_fork(message *msg);
|
|
|
|
/* break.c */
|
|
int do_brk(message *msg);
|
|
int adjust(struct vmproc *rmp, vir_clicks data_clicks, vir_bytes sp);
|
|
int real_brk(struct vmproc *vmp, vir_bytes v);
|
|
|
|
/* signal.c */
|
|
int do_push_sig(message *msg);
|
|
|
|
/* map_mem.c */
|
|
int map_memory(endpoint_t sour, endpoint_t dest, vir_bytes virt_s,
|
|
vir_bytes virt_d, vir_bytes length, int flag);
|
|
int unmap_memory(endpoint_t sour, endpoint_t dest, vir_bytes virt_s,
|
|
vir_bytes virt_d, vir_bytes length, int flag);
|
|
|
|
/* mmap.c */
|
|
int do_mmap(message *msg);
|
|
int do_munmap(message *msg);
|
|
int do_map_phys(message *msg);
|
|
int do_unmap_phys(message *msg);
|
|
int do_remap(message *m);
|
|
int do_get_phys(message *m);
|
|
int do_shared_unmap(message *m);
|
|
int do_get_refcount(message *m);
|
|
|
|
/* pagefaults.c */
|
|
void do_pagefaults(message *m);
|
|
void do_memory(void);
|
|
char *pf_errstr(u32_t err);
|
|
int handle_memory(struct vmproc *vmp, vir_bytes mem, vir_bytes len, int
|
|
wrflag);
|
|
|
|
/* $(ARCH)/pagetable.c */
|
|
void pt_init();
|
|
void pt_init_mem(void);
|
|
void pt_check(struct vmproc *vmp);
|
|
int pt_new(pt_t *pt);
|
|
void pt_free(pt_t *pt);
|
|
int pt_map_in_range(struct vmproc *src_vmp, struct vmproc *dst_vmp,
|
|
vir_bytes start, vir_bytes end);
|
|
int pt_ptmap(struct vmproc *src_vmp, struct vmproc *dst_vmp);
|
|
int pt_ptalloc_in_range(pt_t *pt, vir_bytes start, vir_bytes end, u32_t
|
|
flags, int verify);
|
|
void pt_clearmapcache(void);
|
|
int pt_writemap(struct vmproc * vmp, pt_t *pt, vir_bytes v, phys_bytes
|
|
physaddr, size_t bytes, u32_t flags, u32_t writemapflags);
|
|
int pt_checkrange(pt_t *pt, vir_bytes v, size_t bytes, int write);
|
|
int pt_bind(pt_t *pt, struct vmproc *who);
|
|
void *vm_allocpage(phys_bytes *p, int cat);
|
|
void pt_cycle(void);
|
|
int pt_mapkernel(pt_t *pt);
|
|
void vm_pagelock(void *vir, int lockflag);
|
|
int vm_addrok(void *vir, int write);
|
|
|
|
#if SANITYCHECKS
|
|
void pt_sanitycheck(pt_t *pt, char *file, int line);
|
|
#endif
|
|
|
|
/* slaballoc.c */
|
|
void *slaballoc(int bytes);
|
|
void slabfree(void *mem, int bytes);
|
|
void slabstats(void);
|
|
void slab_sanitycheck(char *file, int line);
|
|
#define SLABALLOC(var) (var = slaballoc(sizeof(*var)))
|
|
#define SLABFREE(ptr) do { slabfree(ptr, sizeof(*(ptr))); (ptr) = NULL; } while(0)
|
|
#if SANITYCHECKS
|
|
|
|
void slabunlock(void *mem, int bytes);
|
|
void slablock(void *mem, int bytes);
|
|
int slabsane_f(char *file, int line, void *mem, int bytes);
|
|
#endif
|
|
|
|
/* region.c */
|
|
void map_region_init(void);
|
|
struct vir_region * map_page_region(struct vmproc *vmp, vir_bytes min,
|
|
vir_bytes max, vir_bytes length, vir_bytes what, u32_t flags, int
|
|
mapflags);
|
|
struct vir_region * map_proc_kernel(struct vmproc *dst);
|
|
int map_region_extend(struct vmproc *vmp, struct vir_region *vr,
|
|
vir_bytes delta);
|
|
int map_region_extend_upto_v(struct vmproc *vmp, vir_bytes vir);
|
|
int map_region_shrink(struct vir_region *vr, vir_bytes delta);
|
|
int map_unmap_region(struct vmproc *vmp, struct vir_region *vr,
|
|
vir_bytes len);
|
|
int map_free_proc(struct vmproc *vmp);
|
|
int map_proc_copy(struct vmproc *dst, struct vmproc *src);
|
|
int map_proc_copy_from(struct vmproc *dst, struct vmproc *src, struct
|
|
vir_region *start_src_vr);
|
|
struct vir_region *map_lookup(struct vmproc *vmp, vir_bytes addr);
|
|
int map_pf(struct vmproc *vmp, struct vir_region *region, vir_bytes
|
|
offset, int write);
|
|
int map_pin_memory(struct vmproc *vmp);
|
|
int map_handle_memory(struct vmproc *vmp, struct vir_region *region,
|
|
vir_bytes offset, vir_bytes len, int write);
|
|
void map_printmap(struct vmproc *vmp);
|
|
int map_writept(struct vmproc *vmp);
|
|
void printregionstats(struct vmproc *vmp);
|
|
phys_bytes map_lookup_phys(struct vmproc *vmp, u32_t tag);
|
|
void map_setparent(struct vmproc *vmp);
|
|
int yielded_block_cmp(struct block_id *, struct block_id *);
|
|
|
|
struct vir_region * map_region_lookup_tag(struct vmproc *vmp, u32_t
|
|
tag);
|
|
void map_region_set_tag(struct vir_region *vr, u32_t tag);
|
|
u32_t map_region_get_tag(struct vir_region *vr);
|
|
int map_remap(struct vmproc *dvmp, vir_bytes da, size_t size, struct
|
|
vir_region *region, vir_bytes *r, int ro);
|
|
int map_get_phys(struct vmproc *vmp, vir_bytes addr, phys_bytes *r);
|
|
int map_get_ref(struct vmproc *vmp, vir_bytes addr, u8_t *cnt);
|
|
|
|
void pb_unreferenced(struct vir_region *region, struct phys_region *pr);
|
|
void get_stats_info(struct vm_stats_info *vsi);
|
|
void get_usage_info(struct vmproc *vmp, struct vm_usage_info *vui);
|
|
int get_region_info(struct vmproc *vmp, struct vm_region_info *vri, int
|
|
count, vir_bytes *nextp);
|
|
int copy_abs2region(phys_bytes abs, struct vir_region *destregion,
|
|
phys_bytes offset, phys_bytes len);
|
|
#if SANITYCHECKS
|
|
void map_sanitycheck(char *file, int line);
|
|
void blockstats(void);
|
|
#endif
|
|
int do_forgetblocks(message *m);
|
|
int do_forgetblock(message *m);
|
|
int do_yieldblockgetblock(message *m);
|
|
vir_bytes free_yielded(vir_bytes bytes);
|
|
|
|
/* rs.c */
|
|
int do_rs_set_priv(message *m);
|
|
int do_rs_update(message *m);
|
|
int do_rs_memctl(message *m);
|
|
|
|
/* queryexit.c */
|
|
int do_query_exit(message *m);
|
|
int do_watch_exit(message *m);
|
|
int do_notify_sig(message *m);
|
|
void init_query_exit(void);
|