| /* |
| * Copyright (C) 2000 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com) |
| * Licensed under the GPL |
| */ |
| |
| #include "linux/percpu.h" |
| #include "asm/pgalloc.h" |
| #include "asm/tlb.h" |
| |
| /* For some reason, mmu_gathers are referenced when CONFIG_SMP is off. */ |
| DEFINE_PER_CPU(struct mmu_gather, mmu_gathers); |
| |
| #ifdef CONFIG_SMP |
| |
| #include "linux/sched.h" |
| #include "linux/module.h" |
| #include "linux/threads.h" |
| #include "linux/interrupt.h" |
| #include "linux/err.h" |
| #include "linux/hardirq.h" |
| #include "asm/smp.h" |
| #include "asm/processor.h" |
| #include "asm/spinlock.h" |
| #include "kern.h" |
| #include "irq_user.h" |
| #include "os.h" |
| |
| /* CPU online map, set by smp_boot_cpus */ |
| cpumask_t cpu_online_map = CPU_MASK_NONE; |
| cpumask_t cpu_possible_map = CPU_MASK_NONE; |
| |
| EXPORT_SYMBOL(cpu_online_map); |
| EXPORT_SYMBOL(cpu_possible_map); |
| |
| /* Per CPU bogomips and other parameters |
| * The only piece used here is the ipi pipe, which is set before SMP is |
| * started and never changed. |
| */ |
| struct cpuinfo_um cpu_data[NR_CPUS]; |
| |
| /* A statistic, can be a little off */ |
| int num_reschedules_sent = 0; |
| |
| /* Not changed after boot */ |
| struct task_struct *idle_threads[NR_CPUS]; |
| |
| void smp_send_reschedule(int cpu) |
| { |
| os_write_file(cpu_data[cpu].ipi_pipe[1], "R", 1); |
| num_reschedules_sent++; |
| } |
| |
| void smp_send_stop(void) |
| { |
| int i; |
| |
| printk(KERN_INFO "Stopping all CPUs..."); |
| for (i = 0; i < num_online_cpus(); i++) { |
| if (i == current_thread->cpu) |
| continue; |
| os_write_file(cpu_data[i].ipi_pipe[1], "S", 1); |
| } |
| printk(KERN_CONT "done\n"); |
| } |
| |
| static cpumask_t smp_commenced_mask = CPU_MASK_NONE; |
| static cpumask_t cpu_callin_map = CPU_MASK_NONE; |
| |
| static int idle_proc(void *cpup) |
| { |
| int cpu = (int) cpup, err; |
| |
| err = os_pipe(cpu_data[cpu].ipi_pipe, 1, 1); |
| if (err < 0) |
| panic("CPU#%d failed to create IPI pipe, err = %d", cpu, -err); |
| |
| os_set_fd_async(cpu_data[cpu].ipi_pipe[0]); |
| |
| wmb(); |
| if (cpu_test_and_set(cpu, cpu_callin_map)) { |
| printk(KERN_ERR "huh, CPU#%d already present??\n", cpu); |
| BUG(); |
| } |
| |
| while (!cpu_isset(cpu, smp_commenced_mask)) |
| cpu_relax(); |
| |
| cpu_set(cpu, cpu_online_map); |
| default_idle(); |
| return 0; |
| } |
| |
| static struct task_struct *idle_thread(int cpu) |
| { |
| struct task_struct *new_task; |
| |
| current->thread.request.u.thread.proc = idle_proc; |
| current->thread.request.u.thread.arg = (void *) cpu; |
| new_task = fork_idle(cpu); |
| if (IS_ERR(new_task)) |
| panic("copy_process failed in idle_thread, error = %ld", |
| PTR_ERR(new_task)); |
| |
| cpu_tasks[cpu] = ((struct cpu_task) |
| { .pid = new_task->thread.mode.tt.extern_pid, |
| .task = new_task } ); |
| idle_threads[cpu] = new_task; |
| panic("skas mode doesn't support SMP"); |
| return new_task; |
| } |
| |
| void smp_prepare_cpus(unsigned int maxcpus) |
| { |
| struct task_struct *idle; |
| unsigned long waittime; |
| int err, cpu, me = smp_processor_id(); |
| int i; |
| |
| for (i = 0; i < ncpus; ++i) |
| cpu_set(i, cpu_possible_map); |
| |
| cpu_clear(me, cpu_online_map); |
| cpu_set(me, cpu_online_map); |
| cpu_set(me, cpu_callin_map); |
| |
| err = os_pipe(cpu_data[me].ipi_pipe, 1, 1); |
| if (err < 0) |
| panic("CPU#0 failed to create IPI pipe, errno = %d", -err); |
| |
| os_set_fd_async(cpu_data[me].ipi_pipe[0]); |
| |
| for (cpu = 1; cpu < ncpus; cpu++) { |
| printk(KERN_INFO "Booting processor %d...\n", cpu); |
| |
| idle = idle_thread(cpu); |
| |
| init_idle(idle, cpu); |
| |
| waittime = 200000000; |
| while (waittime-- && !cpu_isset(cpu, cpu_callin_map)) |
| cpu_relax(); |
| |
| printk(KERN_INFO "%s\n", |
| cpu_isset(cpu, cpu_calling_map) ? "done" : "failed"); |
| } |
| } |
| |
| void smp_prepare_boot_cpu(void) |
| { |
| cpu_set(smp_processor_id(), cpu_online_map); |
| } |
| |
| int __cpu_up(unsigned int cpu) |
| { |
| cpu_set(cpu, smp_commenced_mask); |
| while (!cpu_isset(cpu, cpu_online_map)) |
| mb(); |
| return 0; |
| } |
| |
| int setup_profiling_timer(unsigned int multiplier) |
| { |
| printk(KERN_INFO "setup_profiling_timer\n"); |
| return 0; |
| } |
| |
| void smp_call_function_slave(int cpu); |
| |
| void IPI_handler(int cpu) |
| { |
| unsigned char c; |
| int fd; |
| |
| fd = cpu_data[cpu].ipi_pipe[0]; |
| while (os_read_file(fd, &c, 1) == 1) { |
| switch (c) { |
| case 'C': |
| smp_call_function_slave(cpu); |
| break; |
| |
| case 'R': |
| set_tsk_need_resched(current); |
| break; |
| |
| case 'S': |
| printk(KERN_INFO "CPU#%d stopping\n", cpu); |
| while (1) |
| pause(); |
| break; |
| |
| default: |
| printk(KERN_ERR "CPU#%d received unknown IPI [%c]!\n", |
| cpu, c); |
| break; |
| } |
| } |
| } |
| |
| int hard_smp_processor_id(void) |
| { |
| return pid_to_processor_id(os_getpid()); |
| } |
| |
| static DEFINE_SPINLOCK(call_lock); |
| static atomic_t scf_started; |
| static atomic_t scf_finished; |
| static void (*func)(void *info); |
| static void *info; |
| |
| void smp_call_function_slave(int cpu) |
| { |
| atomic_inc(&scf_started); |
| (*func)(info); |
| atomic_inc(&scf_finished); |
| } |
| |
| int smp_call_function(void (*_func)(void *info), void *_info, int nonatomic, |
| int wait) |
| { |
| int cpus = num_online_cpus() - 1; |
| int i; |
| |
| if (!cpus) |
| return 0; |
| |
| /* Can deadlock when called with interrupts disabled */ |
| WARN_ON(irqs_disabled()); |
| |
| spin_lock_bh(&call_lock); |
| atomic_set(&scf_started, 0); |
| atomic_set(&scf_finished, 0); |
| func = _func; |
| info = _info; |
| |
| for_each_online_cpu(i) |
| os_write_file(cpu_data[i].ipi_pipe[1], "C", 1); |
| |
| while (atomic_read(&scf_started) != cpus) |
| barrier(); |
| |
| if (wait) |
| while (atomic_read(&scf_finished) != cpus) |
| barrier(); |
| |
| spin_unlock_bh(&call_lock); |
| return 0; |
| } |
| |
| #endif |