kernel: spelling fixes
Change-Id: I73c759bdef98be35be77130895ae0ec497e1b954
This commit is contained in:
parent
40fbf55cf3
commit
9f23acf410
23 changed files with 36 additions and 36 deletions
|
@ -151,7 +151,7 @@ void exception_handler(int is_nested, reg_t *saved_lr, int vector)
|
||||||
*/
|
*/
|
||||||
if (is_nested) {
|
if (is_nested) {
|
||||||
/*
|
/*
|
||||||
* if a problem occured while copying a message from userspace because
|
* if a problem occurred while copying a message from userspace because
|
||||||
* of a wrong pointer supplied by userland, handle it the only way we
|
* of a wrong pointer supplied by userland, handle it the only way we
|
||||||
* can handle it ...
|
* can handle it ...
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -60,7 +60,7 @@ typedef int (*kern_phys_map_mapped)(vir_bytes id, vir_bytes new_addr );
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* struct used internally by memory.c to keep a list of
|
* struct used internally by memory.c to keep a list of
|
||||||
* items to map. These should be staticaly allocated
|
* items to map. These should be statically allocated
|
||||||
* in the individual files and passed as argument.
|
* in the individual files and passed as argument.
|
||||||
* The data doesn't need to be initialized. See omap_serial for
|
* The data doesn't need to be initialized. See omap_serial for
|
||||||
* and example usage.
|
* and example usage.
|
||||||
|
@ -108,7 +108,7 @@ int kern_req_phys_map( phys_bytes base_address, vir_bytes io_size,
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Request a physical mapping and put the result in the given prt
|
* Request a physical mapping and put the result in the given prt
|
||||||
* Note that ptr will only be valid once the callback happend.
|
* Note that ptr will only be valid once the callback happened.
|
||||||
*/
|
*/
|
||||||
int kern_phys_map_ptr( phys_bytes base_address, vir_bytes io_size,
|
int kern_phys_map_ptr( phys_bytes base_address, vir_bytes io_size,
|
||||||
kern_phys_map * priv, vir_bytes ptr);
|
kern_phys_map * priv, vir_bytes ptr);
|
||||||
|
|
|
@ -57,7 +57,7 @@ void mem_clear_mapcache(void)
|
||||||
* The target (i.e. in-kernel) mapping area is one of the freepdes[]
|
* The target (i.e. in-kernel) mapping area is one of the freepdes[]
|
||||||
* VM has earlier already told the kernel about that is available. It is
|
* VM has earlier already told the kernel about that is available. It is
|
||||||
* identified as the 'pde' parameter. This value can be chosen freely
|
* identified as the 'pde' parameter. This value can be chosen freely
|
||||||
* by the caller, as long as it is in range (i.e. 0 or higher and corresonds
|
* by the caller, as long as it is in range (i.e. 0 or higher and corresponds
|
||||||
* to a known freepde slot). It is up to the caller to keep track of which
|
* to a known freepde slot). It is up to the caller to keep track of which
|
||||||
* freepde's are in use, and to determine which ones are free to use.
|
* freepde's are in use, and to determine which ones are free to use.
|
||||||
*
|
*
|
||||||
|
@ -902,7 +902,7 @@ int kern_phys_map_mapped_ptr(vir_bytes id, phys_bytes address){
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Request a physical mapping and put the result in the given prt
|
* Request a physical mapping and put the result in the given prt
|
||||||
* Note that ptr will only be valid once the callback happend.
|
* Note that ptr will only be valid once the callback happened.
|
||||||
*/
|
*/
|
||||||
int kern_phys_map_ptr(
|
int kern_phys_map_ptr(
|
||||||
phys_bytes base_address,
|
phys_bytes base_address,
|
||||||
|
|
|
@ -288,7 +288,7 @@ kinfo_t *pre_init(u32_t magic, u32_t ebx)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* pre_init gets executed at the memory location where the kernel was loaded by the boot loader.
|
/* pre_init gets executed at the memory location where the kernel was loaded by the boot loader.
|
||||||
* at that stage we only have a minium set of functionality present (all symbols gets renamed to
|
* at that stage we only have a minimum set of functionality present (all symbols gets renamed to
|
||||||
* ensure this). The following methods are used in that context. Once we jump to kmain they are no
|
* ensure this). The following methods are used in that context. Once we jump to kmain they are no
|
||||||
* longer used and the "real" implementations are visible
|
* longer used and the "real" implementations are visible
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -110,8 +110,8 @@ EXTERN int ioapic_enabled;
|
||||||
struct io_apic {
|
struct io_apic {
|
||||||
unsigned id;
|
unsigned id;
|
||||||
vir_bytes addr; /* presently used address */
|
vir_bytes addr; /* presently used address */
|
||||||
phys_bytes paddr; /* where is it inphys space */
|
phys_bytes paddr; /* where is it in phys space */
|
||||||
vir_bytes vaddr; /* adress after paging s on */
|
vir_bytes vaddr; /* address after paging is on */
|
||||||
unsigned pins;
|
unsigned pins;
|
||||||
unsigned gsi_base;
|
unsigned gsi_base;
|
||||||
};
|
};
|
||||||
|
|
|
@ -59,7 +59,7 @@ void mem_clear_mapcache(void)
|
||||||
* The target (i.e. in-kernel) mapping area is one of the freepdes[]
|
* The target (i.e. in-kernel) mapping area is one of the freepdes[]
|
||||||
* VM has earlier already told the kernel about that is available. It is
|
* VM has earlier already told the kernel about that is available. It is
|
||||||
* identified as the 'pde' parameter. This value can be chosen freely
|
* identified as the 'pde' parameter. This value can be chosen freely
|
||||||
* by the caller, as long as it is in range (i.e. 0 or higher and corresonds
|
* by the caller, as long as it is in range (i.e. 0 or higher and corresponds
|
||||||
* to a known freepde slot). It is up to the caller to keep track of which
|
* to a known freepde slot). It is up to the caller to keep track of which
|
||||||
* freepde's are in use, and to determine which ones are free to use.
|
* freepde's are in use, and to determine which ones are free to use.
|
||||||
*
|
*
|
||||||
|
|
|
@ -39,7 +39,7 @@
|
||||||
movl tmp, STREG(pptr)
|
movl tmp, STREG(pptr)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* restore kernel segments. %cs is aready set and %fs, %gs are not used */
|
* restore kernel segments. %cs is already set and %fs, %gs are not used */
|
||||||
#define RESTORE_KERNEL_SEGS \
|
#define RESTORE_KERNEL_SEGS \
|
||||||
mov $KERN_DS_SELECTOR, %si ;\
|
mov $KERN_DS_SELECTOR, %si ;\
|
||||||
mov %si, %ds ;\
|
mov %si, %ds ;\
|
||||||
|
|
|
@ -10,7 +10,7 @@ int app_cpu_init_timer(unsigned freq);
|
||||||
int timer_int_handler(void);
|
int timer_int_handler(void);
|
||||||
|
|
||||||
int init_local_timer(unsigned freq);
|
int init_local_timer(unsigned freq);
|
||||||
/* sto p the local timer ticking */
|
/* stop the local timer ticking */
|
||||||
void stop_local_timer(void);
|
void stop_local_timer(void);
|
||||||
/* let the time tick again with the original settings after it was stopped */
|
/* let the time tick again with the original settings after it was stopped */
|
||||||
void restart_local_timer(void);
|
void restart_local_timer(void);
|
||||||
|
|
|
@ -54,7 +54,7 @@ EXTERN int verboseflags;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef USE_APIC
|
#ifdef USE_APIC
|
||||||
EXTERN int config_no_apic; /* optionaly turn off apic */
|
EXTERN int config_no_apic; /* optionally turn off apic */
|
||||||
EXTERN int config_apic_timer_x; /* apic timer slowdown factor */
|
EXTERN int config_apic_timer_x; /* apic timer slowdown factor */
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -64,7 +64,7 @@ EXTERN u64_t cpu_hz[CONFIG_MAX_CPUS];
|
||||||
#define cpu_get_freq(cpu) cpu_hz[cpu]
|
#define cpu_get_freq(cpu) cpu_hz[cpu]
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
EXTERN int config_no_smp; /* optionaly turn off SMP */
|
EXTERN int config_no_smp; /* optionally turn off SMP */
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* VM */
|
/* VM */
|
||||||
|
|
|
@ -8,7 +8,7 @@
|
||||||
* rm_irq_handler: deregister an interrupt handler.
|
* rm_irq_handler: deregister an interrupt handler.
|
||||||
* irq_handle: handle a hardware interrupt.
|
* irq_handle: handle a hardware interrupt.
|
||||||
* called by the system dependent part when an
|
* called by the system dependent part when an
|
||||||
* external interrupt occures.
|
* external interrupt occurs.
|
||||||
* enable_irq: enable hook for IRQ.
|
* enable_irq: enable hook for IRQ.
|
||||||
* disable_irq: disable hook for IRQ.
|
* disable_irq: disable hook for IRQ.
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -192,7 +192,7 @@ void kmain(kinfo_t *local_cbi)
|
||||||
/* Assign privilege structure. Force a static privilege id. */
|
/* Assign privilege structure. Force a static privilege id. */
|
||||||
(void) get_priv(rp, static_priv_id(proc_nr));
|
(void) get_priv(rp, static_priv_id(proc_nr));
|
||||||
|
|
||||||
/* Priviliges for kernel tasks. */
|
/* Privileges for kernel tasks. */
|
||||||
if(proc_nr == VM_PROC_NR) {
|
if(proc_nr == VM_PROC_NR) {
|
||||||
priv(rp)->s_flags = VM_F;
|
priv(rp)->s_flags = VM_F;
|
||||||
priv(rp)->s_trap_mask = SRV_T;
|
priv(rp)->s_trap_mask = SRV_T;
|
||||||
|
@ -211,7 +211,7 @@ void kmain(kinfo_t *local_cbi)
|
||||||
ipc_to_m = TSK_M; /* allowed targets */
|
ipc_to_m = TSK_M; /* allowed targets */
|
||||||
kcalls = TSK_KC; /* allowed kernel calls */
|
kcalls = TSK_KC; /* allowed kernel calls */
|
||||||
}
|
}
|
||||||
/* Priviliges for the root system process. */
|
/* Privileges for the root system process. */
|
||||||
else {
|
else {
|
||||||
assert(isrootsysn(proc_nr));
|
assert(isrootsysn(proc_nr));
|
||||||
priv(rp)->s_flags= RSYS_F; /* privilege flags */
|
priv(rp)->s_flags= RSYS_F; /* privilege flags */
|
||||||
|
|
|
@ -117,7 +117,7 @@ void proc_init(void)
|
||||||
struct priv *sp;
|
struct priv *sp;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
/* Clear the process table. Anounce each slot as empty and set up
|
/* Clear the process table. Announce each slot as empty and set up
|
||||||
* mappings for proc_addr() and proc_nr() macros. Do the same for the
|
* mappings for proc_addr() and proc_nr() macros. Do the same for the
|
||||||
* table with privilege structures for the system processes.
|
* table with privilege structures for the system processes.
|
||||||
*/
|
*/
|
||||||
|
@ -175,7 +175,7 @@ static void idle(void)
|
||||||
/* This function is called whenever there is no work to do.
|
/* This function is called whenever there is no work to do.
|
||||||
* Halt the CPU, and measure how many timestamp counter ticks are
|
* Halt the CPU, and measure how many timestamp counter ticks are
|
||||||
* spent not doing anything. This allows test setups to measure
|
* spent not doing anything. This allows test setups to measure
|
||||||
* the CPU utiliziation of certain workloads with high precision.
|
* the CPU utilization of certain workloads with high precision.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
p = get_cpulocal_var(proc_ptr) = get_cpulocal_var_ptr(idle_proc);
|
p = get_cpulocal_var(proc_ptr) = get_cpulocal_var_ptr(idle_proc);
|
||||||
|
@ -637,7 +637,7 @@ endpoint_t src_dst_e; /* src or dst process */
|
||||||
{
|
{
|
||||||
/* Check for deadlock. This can happen if 'caller_ptr' and 'src_dst' have
|
/* Check for deadlock. This can happen if 'caller_ptr' and 'src_dst' have
|
||||||
* a cyclic dependency of blocking send and receive calls. The only cyclic
|
* a cyclic dependency of blocking send and receive calls. The only cyclic
|
||||||
* depency that is not fatal is if the caller and target directly SEND(REC)
|
* dependency that is not fatal is if the caller and target directly SEND(REC)
|
||||||
* and RECEIVE to each other. If a deadlock is found, the group size is
|
* and RECEIVE to each other. If a deadlock is found, the group size is
|
||||||
* returned. Otherwise zero is returned.
|
* returned. Otherwise zero is returned.
|
||||||
*/
|
*/
|
||||||
|
@ -993,7 +993,7 @@ static int mini_receive(struct proc * caller_ptr,
|
||||||
IPC_STATUS_ADD_CALL(caller_ptr, call);
|
IPC_STATUS_ADD_CALL(caller_ptr, call);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* if the message is originaly from the kernel on behalf of this
|
* if the message is originally from the kernel on behalf of this
|
||||||
* process, we must send the status flags accordingly
|
* process, we must send the status flags accordingly
|
||||||
*/
|
*/
|
||||||
if (sender->p_misc_flags & MF_SENDING_FROM_KERNEL) {
|
if (sender->p_misc_flags & MF_SENDING_FROM_KERNEL) {
|
||||||
|
|
|
@ -34,7 +34,7 @@ struct proc {
|
||||||
struct proc *p_scheduler; /* who should get out of quantum msg */
|
struct proc *p_scheduler; /* who should get out of quantum msg */
|
||||||
unsigned p_cpu; /* what CPU is the process running on */
|
unsigned p_cpu; /* what CPU is the process running on */
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
bitchunk_t p_cpu_mask[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* what CPUs is hte
|
bitchunk_t p_cpu_mask[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* what CPUs is the
|
||||||
process allowed to
|
process allowed to
|
||||||
run on */
|
run on */
|
||||||
bitchunk_t p_stale_tlb[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* On which cpu are
|
bitchunk_t p_stale_tlb[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* On which cpu are
|
||||||
|
@ -84,7 +84,7 @@ struct proc {
|
||||||
* memory that isn't present, VM has to fix it. Until it has asked
|
* memory that isn't present, VM has to fix it. Until it has asked
|
||||||
* what needs to be done and fixed it, save necessary state here.
|
* what needs to be done and fixed it, save necessary state here.
|
||||||
*
|
*
|
||||||
* The requestor gets a copy of its request message in reqmsg and gets
|
* The requester gets a copy of its request message in reqmsg and gets
|
||||||
* VMREQUEST set.
|
* VMREQUEST set.
|
||||||
*/
|
*/
|
||||||
struct {
|
struct {
|
||||||
|
|
|
@ -44,7 +44,7 @@ void wait_for_APs_to_finish_booting(void)
|
||||||
BKL_UNLOCK();
|
BKL_UNLOCK();
|
||||||
while (ap_cpus_booted != (n - 1))
|
while (ap_cpus_booted != (n - 1))
|
||||||
arch_pause();
|
arch_pause();
|
||||||
/* now we have to take the lock again as we continu execution */
|
/* now we have to take the lock again as we continue execution */
|
||||||
BKL_LOCK();
|
BKL_LOCK();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -13,7 +13,7 @@
|
||||||
EXTERN unsigned ncpus;
|
EXTERN unsigned ncpus;
|
||||||
/* Number of virtual strands per physical core */
|
/* Number of virtual strands per physical core */
|
||||||
EXTERN unsigned ht_per_core;
|
EXTERN unsigned ht_per_core;
|
||||||
/* which cpu is bootstraping */
|
/* which cpu is bootstrapping */
|
||||||
EXTERN unsigned bsp_cpu_id;
|
EXTERN unsigned bsp_cpu_id;
|
||||||
|
|
||||||
#define cpu_is_bsp(cpu) (bsp_cpu_id == cpu)
|
#define cpu_is_bsp(cpu) (bsp_cpu_id == cpu)
|
||||||
|
@ -62,7 +62,7 @@ void smp_ipi_sched_handler(void);
|
||||||
void smp_schedule(unsigned cpu);
|
void smp_schedule(unsigned cpu);
|
||||||
/* stop a processes on a different cpu */
|
/* stop a processes on a different cpu */
|
||||||
void smp_schedule_stop_proc(struct proc * p);
|
void smp_schedule_stop_proc(struct proc * p);
|
||||||
/* stop a process on a different cpu because its adress space is being changed */
|
/* stop a process on a different cpu because its address space is being changed */
|
||||||
void smp_schedule_vminhibit(struct proc * p);
|
void smp_schedule_vminhibit(struct proc * p);
|
||||||
/* stop the process and for saving its full context */
|
/* stop the process and for saving its full context */
|
||||||
void smp_schedule_stop_proc_save_ctx(struct proc * p);
|
void smp_schedule_stop_proc_save_ctx(struct proc * p);
|
||||||
|
|
|
@ -180,7 +180,7 @@ void system_init(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Initialize the call vector to a safe default handler. Some system calls
|
/* Initialize the call vector to a safe default handler. Some system calls
|
||||||
* may be disabled or nonexistant. Then explicitely map known calls to their
|
* may be disabled or nonexistant. Then explicitly map known calls to their
|
||||||
* handler functions. This is done with a macro that gives a compile error
|
* handler functions. This is done with a macro that gives a compile error
|
||||||
* if an illegal call number is used. The ordering is not important here.
|
* if an illegal call number is used. The ordering is not important here.
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -58,7 +58,7 @@ int do_copy(struct proc * caller, message * m_ptr)
|
||||||
*/
|
*/
|
||||||
for (i=_SRC_; i<=_DST_; i++) {
|
for (i=_SRC_; i<=_DST_; i++) {
|
||||||
int p;
|
int p;
|
||||||
/* Check if process number was given implictly with SELF and is valid. */
|
/* Check if process number was given implicitly with SELF and is valid. */
|
||||||
if (vir_addr[i].proc_nr_e == SELF)
|
if (vir_addr[i].proc_nr_e == SELF)
|
||||||
vir_addr[i].proc_nr_e = caller->p_endpoint;
|
vir_addr[i].proc_nr_e = caller->p_endpoint;
|
||||||
if (vir_addr[i].proc_nr_e != NONE) {
|
if (vir_addr[i].proc_nr_e != NONE) {
|
||||||
|
|
|
@ -94,7 +94,7 @@ int do_fork(struct proc * caller, message * m_ptr)
|
||||||
|
|
||||||
/* If the parent is a privileged process, take away the privileges from the
|
/* If the parent is a privileged process, take away the privileges from the
|
||||||
* child process and inhibit it from running by setting the NO_PRIV flag.
|
* child process and inhibit it from running by setting the NO_PRIV flag.
|
||||||
* The caller should explicitely set the new privileges before executing.
|
* The caller should explicitly set the new privileges before executing.
|
||||||
*/
|
*/
|
||||||
if (priv(rpp)->s_flags & SYS_PROC) {
|
if (priv(rpp)->s_flags & SYS_PROC) {
|
||||||
rpc->p_priv = priv_addr(USER_PRIV_ID);
|
rpc->p_priv = priv_addr(USER_PRIV_ID);
|
||||||
|
|
|
@ -38,7 +38,7 @@ int do_privctl(struct proc * caller, message * m_ptr)
|
||||||
struct priv priv;
|
struct priv priv;
|
||||||
int irq;
|
int irq;
|
||||||
|
|
||||||
/* Check whether caller is allowed to make this call. Privileged proceses
|
/* Check whether caller is allowed to make this call. Privileged processes
|
||||||
* can only update the privileges of processes that are inhibited from
|
* can only update the privileges of processes that are inhibited from
|
||||||
* running by the RTS_NO_PRIV flag. This flag is set when a privileged process
|
* running by the RTS_NO_PRIV flag. This flag is set when a privileged process
|
||||||
* forks.
|
* forks.
|
||||||
|
@ -221,7 +221,7 @@ int do_privctl(struct proc * caller, message * m_ptr)
|
||||||
return r;
|
return r;
|
||||||
priv(rp)->s_flags |= CHECK_MEM; /* Check memory mappings */
|
priv(rp)->s_flags |= CHECK_MEM; /* Check memory mappings */
|
||||||
|
|
||||||
/* When restarting a driver, check if it already has the premission */
|
/* When restarting a driver, check if it already has the permission */
|
||||||
for (i = 0; i < priv(rp)->s_nr_mem_range; i++) {
|
for (i = 0; i < priv(rp)->s_nr_mem_range; i++) {
|
||||||
if (priv(rp)->s_mem_tab[i].mr_base == mem_range.mr_base &&
|
if (priv(rp)->s_mem_tab[i].mr_base == mem_range.mr_base &&
|
||||||
priv(rp)->s_mem_tab[i].mr_limit == mem_range.mr_limit)
|
priv(rp)->s_mem_tab[i].mr_limit == mem_range.mr_limit)
|
||||||
|
@ -253,7 +253,7 @@ int do_privctl(struct proc * caller, message * m_ptr)
|
||||||
KERNEL, (vir_bytes) &irq, sizeof(irq));
|
KERNEL, (vir_bytes) &irq, sizeof(irq));
|
||||||
priv(rp)->s_flags |= CHECK_IRQ; /* Check IRQs */
|
priv(rp)->s_flags |= CHECK_IRQ; /* Check IRQs */
|
||||||
|
|
||||||
/* When restarting a driver, check if it already has the premission */
|
/* When restarting a driver, check if it already has the permission */
|
||||||
for (i = 0; i < priv(rp)->s_nr_irq; i++) {
|
for (i = 0; i < priv(rp)->s_nr_irq; i++) {
|
||||||
if (priv(rp)->s_irq_tab[i] == irq)
|
if (priv(rp)->s_irq_tab[i] == irq)
|
||||||
return OK;
|
return OK;
|
||||||
|
|
|
@ -28,7 +28,7 @@ int do_vdevio(struct proc * caller, message * m_ptr)
|
||||||
/* Perform a series of device I/O on behalf of a non-kernel process. The
|
/* Perform a series of device I/O on behalf of a non-kernel process. The
|
||||||
* I/O addresses and I/O values are fetched from and returned to some buffer
|
* I/O addresses and I/O values are fetched from and returned to some buffer
|
||||||
* in user space. The actual I/O is wrapped by lock() and unlock() to prevent
|
* in user space. The actual I/O is wrapped by lock() and unlock() to prevent
|
||||||
* that I/O batch from being interrrupted.
|
* that I/O batch from being interrupted.
|
||||||
* This is the counterpart of do_devio, which performs a single device I/O.
|
* This is the counterpart of do_devio, which performs a single device I/O.
|
||||||
*/
|
*/
|
||||||
int vec_size; /* size of vector */
|
int vec_size; /* size of vector */
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
/*
|
/*
|
||||||
* This is arch independent NMI watchdog implementaion part. It is used to
|
* This is arch independent NMI watchdog implementation part. It is used to
|
||||||
* detect kernel lockups and help debugging. each architecture must add its own
|
* detect kernel lockups and help debugging. each architecture must add its own
|
||||||
* low level code that triggers periodic checks
|
* low level code that triggers periodic checks
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -35,9 +35,9 @@ void arch_watchdog_stop(void);
|
||||||
/* if the watchdog detects lockup, let the arch code to handle it */
|
/* if the watchdog detects lockup, let the arch code to handle it */
|
||||||
void arch_watchdog_lockup(const struct nmi_frame * frame);
|
void arch_watchdog_lockup(const struct nmi_frame * frame);
|
||||||
|
|
||||||
/* generic NMI handler. Takes one agument which points to where the arch
|
/* generic NMI handler. Takes one argument which points to where the arch
|
||||||
* specific low level handler dumped CPU information and can be inspected by the
|
* specific low level handler dumped CPU information and can be inspected by the
|
||||||
* arch specific code of the watchdog implementaion */
|
* arch specific code of the watchdog implementation */
|
||||||
void nmi_watchdog_handler(struct nmi_frame * frame);
|
void nmi_watchdog_handler(struct nmi_frame * frame);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -30,7 +30,7 @@ EXTERN struct schedproc {
|
||||||
unsigned priority; /* the process' current priority */
|
unsigned priority; /* the process' current priority */
|
||||||
unsigned time_slice; /* this process's time slice */
|
unsigned time_slice; /* this process's time slice */
|
||||||
unsigned cpu; /* what CPU is the process running on */
|
unsigned cpu; /* what CPU is the process running on */
|
||||||
bitchunk_t cpu_mask[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* what CPUs is hte
|
bitchunk_t cpu_mask[BITMAP_CHUNKS(CONFIG_MAX_CPUS)]; /* what CPUs is the
|
||||||
process allowed
|
process allowed
|
||||||
to run on */
|
to run on */
|
||||||
} schedproc[NR_PROCS];
|
} schedproc[NR_PROCS];
|
||||||
|
|
Loading…
Reference in a new issue