| /* |
| * include/asm-alpha/processor.h |
| * |
| * Copyright (C) 1994 Linus Torvalds |
| */ |
| |
| #ifndef __ASM_ALPHA_PROCESSOR_H |
| #define __ASM_ALPHA_PROCESSOR_H |
| |
| #include <linux/personality.h> /* for ADDR_LIMIT_32BIT */ |
| |
| /* |
| * Returns current instruction pointer ("program counter"). |
| */ |
| #define current_text_addr() \ |
| ({ void *__pc; __asm__ ("br %0,.+4" : "=r"(__pc)); __pc; }) |
| |
| /* |
| * We have a 42-bit user address space: 4TB user VM... |
| */ |
| #define TASK_SIZE (0x40000000000UL) |
| |
| /* This decides where the kernel will search for a free chunk of vm |
| * space during mmap's. |
| */ |
| #define TASK_UNMAPPED_BASE \ |
| ((current->personality & ADDR_LIMIT_32BIT) ? 0x40000000 : TASK_SIZE / 2) |
| |
| typedef struct { |
| unsigned long seg; |
| } mm_segment_t; |
| |
| /* This is dead. Everything has been moved to thread_info. */ |
| struct thread_struct { }; |
| #define INIT_THREAD { } |
| |
| /* Return saved PC of a blocked thread. */ |
| struct task_struct; |
| extern unsigned long thread_saved_pc(struct task_struct *); |
| |
| /* Do necessary setup to start up a newly executed thread. */ |
| extern void start_thread(struct pt_regs *, unsigned long, unsigned long); |
| |
| /* Free all resources held by a thread. */ |
| extern void release_thread(struct task_struct *); |
| |
| /* Prepare to copy thread state - unlazy all lazy status */ |
| #define prepare_to_copy(tsk) do { } while (0) |
| |
| /* Create a kernel thread without removing it from tasklists. */ |
| extern long kernel_thread(int (*fn)(void *), void *arg, unsigned long flags); |
| |
| unsigned long get_wchan(struct task_struct *p); |
| |
| #define KSTK_EIP(tsk) (task_pt_regs(tsk)->pc) |
| |
| #define KSTK_ESP(tsk) \ |
| ((tsk) == current ? rdusp() : task_thread_info(tsk)->pcb.usp) |
| |
| #define cpu_relax() barrier() |
| |
| #define ARCH_HAS_PREFETCH |
| #define ARCH_HAS_PREFETCHW |
| #define ARCH_HAS_SPINLOCK_PREFETCH |
| |
| #ifndef CONFIG_SMP |
| /* Nothing to prefetch. */ |
| #define spin_lock_prefetch(lock) do { } while (0) |
| #endif |
| |
| extern inline void prefetch(const void *ptr) |
| { |
| __builtin_prefetch(ptr, 0, 3); |
| } |
| |
| extern inline void prefetchw(const void *ptr) |
| { |
| __builtin_prefetch(ptr, 1, 3); |
| } |
| |
| #ifdef CONFIG_SMP |
| extern inline void spin_lock_prefetch(const void *ptr) |
| { |
| __builtin_prefetch(ptr, 1, 3); |
| } |
| #endif |
| |
| #endif /* __ASM_ALPHA_PROCESSOR_H */ |