| /* |
| * Copyright (C) 2002 Jeff Dike (jdike@karaya.com) |
| * Licensed under the GPL |
| */ |
| |
| #include "linux/sched.h" |
| #include "linux/slab.h" |
| #include "linux/ptrace.h" |
| #include "linux/proc_fs.h" |
| #include "linux/file.h" |
| #include "linux/errno.h" |
| #include "linux/init.h" |
| #include "asm/uaccess.h" |
| #include "asm/atomic.h" |
| #include "kern_util.h" |
| #include "time_user.h" |
| #include "signal_user.h" |
| #include "skas.h" |
| #include "os.h" |
| #include "user_util.h" |
| #include "tlb.h" |
| #include "kern.h" |
| #include "mode.h" |
| #include "proc_mm.h" |
| #include "registers.h" |
| |
| void switch_to_skas(void *prev, void *next) |
| { |
| struct task_struct *from, *to; |
| |
| from = prev; |
| to = next; |
| |
| /* XXX need to check runqueues[cpu].idle */ |
| if(current->pid == 0) |
| switch_timers(0); |
| |
| switch_threads(&from->thread.mode.skas.switch_buf, |
| to->thread.mode.skas.switch_buf); |
| |
| if(current->pid == 0) |
| switch_timers(1); |
| } |
| |
| extern void schedule_tail(struct task_struct *prev); |
| |
| void new_thread_handler(int sig) |
| { |
| int (*fn)(void *), n; |
| void *arg; |
| |
| fn = current->thread.request.u.thread.proc; |
| arg = current->thread.request.u.thread.arg; |
| change_sig(SIGUSR1, 1); |
| thread_wait(¤t->thread.mode.skas.switch_buf, |
| current->thread.mode.skas.fork_buf); |
| |
| if(current->thread.prev_sched != NULL) |
| schedule_tail(current->thread.prev_sched); |
| current->thread.prev_sched = NULL; |
| |
| /* The return value is 1 if the kernel thread execs a process, |
| * 0 if it just exits |
| */ |
| n = run_kernel_thread(fn, arg, ¤t->thread.exec_buf); |
| if(n == 1){ |
| /* Handle any immediate reschedules or signals */ |
| interrupt_end(); |
| userspace(¤t->thread.regs.regs); |
| } |
| else do_exit(0); |
| } |
| |
| void new_thread_proc(void *stack, void (*handler)(int sig)) |
| { |
| init_new_thread_stack(stack, handler); |
| os_usr1_process(os_getpid()); |
| } |
| |
| void release_thread_skas(struct task_struct *task) |
| { |
| } |
| |
| void fork_handler(int sig) |
| { |
| change_sig(SIGUSR1, 1); |
| thread_wait(¤t->thread.mode.skas.switch_buf, |
| current->thread.mode.skas.fork_buf); |
| |
| force_flush_all(); |
| if(current->thread.prev_sched == NULL) |
| panic("blech"); |
| |
| schedule_tail(current->thread.prev_sched); |
| current->thread.prev_sched = NULL; |
| |
| /* Handle any immediate reschedules or signals */ |
| interrupt_end(); |
| userspace(¤t->thread.regs.regs); |
| } |
| |
| int copy_thread_skas(int nr, unsigned long clone_flags, unsigned long sp, |
| unsigned long stack_top, struct task_struct * p, |
| struct pt_regs *regs) |
| { |
| void (*handler)(int); |
| |
| if(current->thread.forking){ |
| memcpy(&p->thread.regs.regs.skas, ®s->regs.skas, |
| sizeof(p->thread.regs.regs.skas)); |
| REGS_SET_SYSCALL_RETURN(p->thread.regs.regs.skas.regs, 0); |
| if(sp != 0) REGS_SP(p->thread.regs.regs.skas.regs) = sp; |
| |
| handler = fork_handler; |
| } |
| else { |
| init_thread_registers(&p->thread.regs.regs); |
| p->thread.request.u.thread = current->thread.request.u.thread; |
| handler = new_thread_handler; |
| } |
| |
| new_thread(p->thread_info, &p->thread.mode.skas.switch_buf, |
| &p->thread.mode.skas.fork_buf, handler); |
| return(0); |
| } |
| |
| extern void map_stub_pages(int fd, unsigned long code, |
| unsigned long data, unsigned long stack); |
| int new_mm(int from, unsigned long stack) |
| { |
| struct proc_mm_op copy; |
| int n, fd; |
| |
| fd = os_open_file("/proc/mm", of_cloexec(of_write(OPENFLAGS())), 0); |
| if(fd < 0) |
| return(fd); |
| |
| if(from != -1){ |
| copy = ((struct proc_mm_op) { .op = MM_COPY_SEGMENTS, |
| .u = |
| { .copy_segments = from } } ); |
| n = os_write_file(fd, ©, sizeof(copy)); |
| if(n != sizeof(copy)) |
| printk("new_mm : /proc/mm copy_segments failed, " |
| "err = %d\n", -n); |
| } |
| |
| if(!ptrace_faultinfo) |
| map_stub_pages(fd, CONFIG_STUB_CODE, CONFIG_STUB_DATA, stack); |
| |
| return(fd); |
| } |
| |
| void init_idle_skas(void) |
| { |
| cpu_tasks[current_thread->cpu].pid = os_getpid(); |
| default_idle(); |
| } |
| |
| extern void start_kernel(void); |
| |
| static int start_kernel_proc(void *unused) |
| { |
| int pid; |
| |
| block_signals(); |
| pid = os_getpid(); |
| |
| cpu_tasks[0].pid = pid; |
| cpu_tasks[0].task = current; |
| #ifdef CONFIG_SMP |
| cpu_online_map = cpumask_of_cpu(0); |
| #endif |
| start_kernel(); |
| return(0); |
| } |
| |
| extern int userspace_pid[]; |
| |
| int start_uml_skas(void) |
| { |
| if(proc_mm) |
| userspace_pid[0] = start_userspace(0); |
| |
| init_new_thread_signals(1); |
| |
| init_task.thread.request.u.thread.proc = start_kernel_proc; |
| init_task.thread.request.u.thread.arg = NULL; |
| return(start_idle_thread(init_task.thread_info, |
| &init_task.thread.mode.skas.switch_buf, |
| &init_task.thread.mode.skas.fork_buf)); |
| } |
| |
| int external_pid_skas(struct task_struct *task) |
| { |
| #warning Need to look up userspace_pid by cpu |
| return(userspace_pid[0]); |
| } |
| |
| int thread_pid_skas(struct task_struct *task) |
| { |
| #warning Need to look up userspace_pid by cpu |
| return(userspace_pid[0]); |
| } |
| |
| void kill_off_processes_skas(void) |
| { |
| if(proc_mm) |
| #warning need to loop over userspace_pids in kill_off_processes_skas |
| os_kill_ptraced_process(userspace_pid[0], 1); |
| else { |
| struct task_struct *p; |
| int pid, me; |
| |
| me = os_getpid(); |
| for_each_process(p){ |
| if(p->mm == NULL) |
| continue; |
| |
| pid = p->mm->context.skas.id.u.pid; |
| os_kill_ptraced_process(pid, 1); |
| } |
| } |
| } |
| |
| unsigned long current_stub_stack(void) |
| { |
| if(current->mm == NULL) |
| return(0); |
| |
| return(current->mm->context.skas.id.stack); |
| } |