blob: 83719eee2d134526c510be235d406df45fbe2106 [file] [log] [blame]
#ifndef ASM_EDAC_H
#define ASM_EDAC_H
/* ECC atomic, DMA, SMP and interrupt safe scrub function */
static inline void atomic_scrub(void *va, u32 size)
{
unsigned long *virt_addr = va;
unsigned long temp;
u32 i;
for (i = 0; i < size / sizeof(unsigned long); i++, virt_addr++) {
/*
* Very carefully read and write to memory atomically
* so we are interrupt, DMA and SMP safe.
*
* Intel: asm("lock; addl $0, %0"::"m"(*virt_addr));
*/
__asm__ __volatile__ (
" .set mips3 \n"
"1: ll %0, %1 # atomic_add \n"
" ll %0, %1 # atomic_add \n"
" addu %0, $0 \n"
" sc %0, %1 \n"
" beqz %0, 1b \n"
" .set mips0 \n"
: "=&r" (temp), "=m" (*virt_addr)
: "m" (*virt_addr));
}
}
#endif