/* SMP global caching code * * Copyright (C) 2010 Red Hat, Inc. All Rights Reserved. * Written by David Howells (dhowells@redhat.com) * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public Licence * as published by the Free Software Foundation; either version * 2 of the Licence, or (at your option) any later version. */ #include <linux/module.h> #include <linux/mm.h> #include <linux/mman.h> #include <linux/threads.h> #include <linux/interrupt.h> #include <asm/page.h> #include <asm/pgtable.h> #include <asm/processor.h> #include <asm/cacheflush.h> #include <asm/io.h> #include <asm/uaccess.h> #include <asm/smp.h> #include "cache-smp.h" DEFINE_SPINLOCK(smp_cache_lock); static unsigned long smp_cache_mask; static unsigned long smp_cache_start; static unsigned long smp_cache_end; static cpumask_t smp_cache_ipi_map; /* Bitmask of cache IPI done CPUs */ /** * smp_cache_interrupt - Handle IPI request to flush caches. * * Handle a request delivered by IPI to flush the current CPU's * caches. The parameters are stored in smp_cache_*. */ void smp_cache_interrupt(void) { unsigned long opr_mask = smp_cache_mask; switch ((enum smp_dcache_ops)(opr_mask & SMP_DCACHE_OP_MASK)) { case SMP_DCACHE_NOP: break; case SMP_DCACHE_INV: mn10300_local_dcache_inv(); break; case SMP_DCACHE_INV_RANGE: mn10300_local_dcache_inv_range(smp_cache_start, smp_cache_end); break; case SMP_DCACHE_FLUSH: mn10300_local_dcache_flush(); break; case SMP_DCACHE_FLUSH_RANGE: mn10300_local_dcache_flush_range(smp_cache_start, smp_cache_end); break; case SMP_DCACHE_FLUSH_INV: mn10300_local_dcache_flush_inv(); break; case SMP_DCACHE_FLUSH_INV_RANGE: mn10300_local_dcache_flush_inv_range(smp_cache_start, smp_cache_end); break; } switch ((enum smp_icache_ops)(opr_mask & SMP_ICACHE_OP_MASK)) { case SMP_ICACHE_NOP: break; case SMP_ICACHE_INV: mn10300_local_icache_inv(); break; case SMP_ICACHE_INV_RANGE: mn10300_local_icache_inv_range(smp_cache_start, smp_cache_end); break; } cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map); } /** * smp_cache_call - Issue an IPI to request the other CPUs flush caches * @opr_mask: Cache operation flags * @start: Start address of request * @end: End address of request * * Send cache flush IPI to other CPUs. This invokes smp_cache_interrupt() * above on those other CPUs and then waits for them to finish. * * The caller must hold smp_cache_lock. */ void smp_cache_call(unsigned long opr_mask, unsigned long start, unsigned long end) { smp_cache_mask = opr_mask; smp_cache_start = start; smp_cache_end = end; cpumask_copy(&smp_cache_ipi_map, cpu_online_mask); cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map); send_IPI_allbutself(FLUSH_CACHE_IPI); while (!cpumask_empty(&smp_cache_ipi_map)) /* nothing. lockup detection does not belong here */ mb(); }