2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
|
|
* for more details.
|
|
|
|
*
|
2011-04-04 21:15:29 +00:00
|
|
|
* Copyright (C) 1996 David S. Miller (davem@davemloft.net)
|
2005-04-16 22:20:36 +00:00
|
|
|
* Copyright (C) 1997, 1998, 1999, 2000 Ralf Baechle ralf@gnu.org
|
|
|
|
* Carsten Langgaard, carstenl@mips.com
|
|
|
|
* Copyright (C) 2002 MIPS Technologies, Inc. All rights reserved.
|
|
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/sched.h>
|
2009-06-19 13:05:26 +00:00
|
|
|
#include <linux/smp.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/mm.h>
|
2009-05-28 00:47:44 +00:00
|
|
|
#include <linux/hugetlb.h>
|
2012-11-22 02:34:10 +00:00
|
|
|
#include <linux/module.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
#include <asm/cpu.h>
|
|
|
|
#include <asm/bootinfo.h>
|
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
#include <asm/pgtable.h>
|
2011-11-28 16:11:28 +00:00
|
|
|
#include <asm/tlbmisc.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
extern void build_tlb_refill_handler(void);
|
|
|
|
|
2005-04-02 10:21:56 +00:00
|
|
|
/*
|
|
|
|
* Make sure all entries differ. If they're not different
|
|
|
|
* MIPS32 will take revenge ...
|
|
|
|
*/
|
|
|
|
#define UNIQUE_ENTRYHI(idx) (CKSEG0 + ((idx) << (PAGE_SHIFT + 1)))
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
/* Atomicity and interruptability */
|
|
|
|
#ifdef CONFIG_MIPS_MT_SMTC
|
|
|
|
|
|
|
|
#include <asm/smtc.h>
|
|
|
|
#include <asm/mipsmtregs.h>
|
|
|
|
|
|
|
|
#define ENTER_CRITICAL(flags) \
|
|
|
|
{ \
|
|
|
|
unsigned int mvpflags; \
|
|
|
|
local_irq_save(flags);\
|
|
|
|
mvpflags = dvpe()
|
|
|
|
#define EXIT_CRITICAL(flags) \
|
|
|
|
evpe(mvpflags); \
|
|
|
|
local_irq_restore(flags); \
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
|
|
|
|
#define ENTER_CRITICAL(flags) local_irq_save(flags)
|
|
|
|
#define EXIT_CRITICAL(flags) local_irq_restore(flags)
|
|
|
|
|
|
|
|
#endif /* CONFIG_MIPS_MT_SMTC */
|
|
|
|
|
2007-06-06 06:52:43 +00:00
|
|
|
#if defined(CONFIG_CPU_LOONGSON2)
|
|
|
|
/*
|
|
|
|
* LOONGSON2 has a 4 entry itlb which is a subset of dtlb,
|
|
|
|
* unfortrunately, itlb is not totally transparent to software.
|
|
|
|
*/
|
|
|
|
#define FLUSH_ITLB write_c0_diag(4);
|
|
|
|
|
|
|
|
#define FLUSH_ITLB_VM(vma) { if ((vma)->vm_flags & VM_EXEC) write_c0_diag(4); }
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
|
|
|
#define FLUSH_ITLB
|
|
|
|
#define FLUSH_ITLB_VM(vma)
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
void local_flush_tlb_all(void)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
unsigned long old_ctx;
|
|
|
|
int entry;
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Save old context and create impossible VPN2 value */
|
|
|
|
old_ctx = read_c0_entryhi();
|
|
|
|
write_c0_entrylo0(0);
|
|
|
|
write_c0_entrylo1(0);
|
|
|
|
|
|
|
|
entry = read_c0_wired();
|
|
|
|
|
|
|
|
/* Blast 'em all away. */
|
|
|
|
while (entry < current_cpu_data.tlbsize) {
|
2005-04-02 10:21:56 +00:00
|
|
|
/* Make sure all entries differ. */
|
|
|
|
write_c0_entryhi(UNIQUE_ENTRYHI(entry));
|
2005-04-16 22:20:36 +00:00
|
|
|
write_c0_index(entry);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
entry++;
|
|
|
|
}
|
|
|
|
tlbw_use_hazard();
|
|
|
|
write_c0_entryhi(old_ctx);
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB;
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2012-11-22 02:34:10 +00:00
|
|
|
EXPORT_SYMBOL(local_flush_tlb_all);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-04-02 10:21:56 +00:00
|
|
|
/* All entries common to a mm share an asid. To effectively flush
|
|
|
|
these entries, we just bump the asid. */
|
2005-04-16 22:20:36 +00:00
|
|
|
void local_flush_tlb_mm(struct mm_struct *mm)
|
|
|
|
{
|
2005-04-02 10:21:56 +00:00
|
|
|
int cpu;
|
|
|
|
|
|
|
|
preempt_disable();
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-04-02 10:21:56 +00:00
|
|
|
cpu = smp_processor_id();
|
|
|
|
|
|
|
|
if (cpu_context(cpu, mm) != 0) {
|
|
|
|
drop_mmu_context(mm, cpu);
|
|
|
|
}
|
|
|
|
|
|
|
|
preempt_enable();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
|
|
|
unsigned long end)
|
|
|
|
{
|
|
|
|
struct mm_struct *mm = vma->vm_mm;
|
|
|
|
int cpu = smp_processor_id();
|
|
|
|
|
|
|
|
if (cpu_context(cpu, mm) != 0) {
|
2009-05-20 06:12:32 +00:00
|
|
|
unsigned long size, flags;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2012-12-03 20:44:26 +00:00
|
|
|
start = round_down(start, PAGE_SIZE << 1);
|
|
|
|
end = round_up(end, PAGE_SIZE << 1);
|
|
|
|
size = (end - start) >> (PAGE_SHIFT + 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (size <= current_cpu_data.tlbsize/2) {
|
|
|
|
int oldpid = read_c0_entryhi();
|
|
|
|
int newpid = cpu_asid(cpu, mm);
|
|
|
|
|
|
|
|
while (start < end) {
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
write_c0_entryhi(start | newpid);
|
2012-12-03 20:44:26 +00:00
|
|
|
start += (PAGE_SIZE << 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_probe();
|
2006-09-08 02:16:21 +00:00
|
|
|
tlb_probe_hazard();
|
2005-04-16 22:20:36 +00:00
|
|
|
idx = read_c0_index();
|
|
|
|
write_c0_entrylo0(0);
|
|
|
|
write_c0_entrylo1(0);
|
|
|
|
if (idx < 0)
|
|
|
|
continue;
|
|
|
|
/* Make sure all entries differ. */
|
2005-04-02 10:21:56 +00:00
|
|
|
write_c0_entryhi(UNIQUE_ENTRYHI(idx));
|
2005-04-16 22:20:36 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
}
|
|
|
|
tlbw_use_hazard();
|
|
|
|
write_c0_entryhi(oldpid);
|
|
|
|
} else {
|
|
|
|
drop_mmu_context(mm, cpu);
|
|
|
|
}
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB;
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void local_flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
|
|
{
|
2009-05-20 06:12:32 +00:00
|
|
|
unsigned long size, flags;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
|
|
|
|
size = (size + 1) >> 1;
|
|
|
|
if (size <= current_cpu_data.tlbsize / 2) {
|
|
|
|
int pid = read_c0_entryhi();
|
|
|
|
|
|
|
|
start &= (PAGE_MASK << 1);
|
|
|
|
end += ((PAGE_SIZE << 1) - 1);
|
|
|
|
end &= (PAGE_MASK << 1);
|
|
|
|
|
|
|
|
while (start < end) {
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
write_c0_entryhi(start);
|
|
|
|
start += (PAGE_SIZE << 1);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_probe();
|
2006-09-08 02:16:21 +00:00
|
|
|
tlb_probe_hazard();
|
2005-04-16 22:20:36 +00:00
|
|
|
idx = read_c0_index();
|
|
|
|
write_c0_entrylo0(0);
|
|
|
|
write_c0_entrylo1(0);
|
|
|
|
if (idx < 0)
|
|
|
|
continue;
|
|
|
|
/* Make sure all entries differ. */
|
2005-04-02 10:21:56 +00:00
|
|
|
write_c0_entryhi(UNIQUE_ENTRYHI(idx));
|
2005-04-16 22:20:36 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
}
|
|
|
|
tlbw_use_hazard();
|
|
|
|
write_c0_entryhi(pid);
|
|
|
|
} else {
|
|
|
|
local_flush_tlb_all();
|
|
|
|
}
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB;
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page)
|
|
|
|
{
|
|
|
|
int cpu = smp_processor_id();
|
|
|
|
|
|
|
|
if (cpu_context(cpu, vma->vm_mm) != 0) {
|
|
|
|
unsigned long flags;
|
|
|
|
int oldpid, newpid, idx;
|
|
|
|
|
|
|
|
newpid = cpu_asid(cpu, vma->vm_mm);
|
|
|
|
page &= (PAGE_MASK << 1);
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
oldpid = read_c0_entryhi();
|
|
|
|
write_c0_entryhi(page | newpid);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_probe();
|
2006-09-08 02:16:21 +00:00
|
|
|
tlb_probe_hazard();
|
2005-04-16 22:20:36 +00:00
|
|
|
idx = read_c0_index();
|
|
|
|
write_c0_entrylo0(0);
|
|
|
|
write_c0_entrylo1(0);
|
|
|
|
if (idx < 0)
|
|
|
|
goto finish;
|
|
|
|
/* Make sure all entries differ. */
|
2005-04-02 10:21:56 +00:00
|
|
|
write_c0_entryhi(UNIQUE_ENTRYHI(idx));
|
2005-04-16 22:20:36 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
tlbw_use_hazard();
|
|
|
|
|
|
|
|
finish:
|
|
|
|
write_c0_entryhi(oldpid);
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB_VM(vma);
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This one is only used for pages with the global bit set so we don't care
|
|
|
|
* much about the ASID.
|
|
|
|
*/
|
|
|
|
void local_flush_tlb_one(unsigned long page)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
int oldpid, idx;
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
oldpid = read_c0_entryhi();
|
2005-04-02 10:21:56 +00:00
|
|
|
page &= (PAGE_MASK << 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
write_c0_entryhi(page);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_probe();
|
2006-09-08 02:16:21 +00:00
|
|
|
tlb_probe_hazard();
|
2005-04-16 22:20:36 +00:00
|
|
|
idx = read_c0_index();
|
|
|
|
write_c0_entrylo0(0);
|
|
|
|
write_c0_entrylo1(0);
|
|
|
|
if (idx >= 0) {
|
|
|
|
/* Make sure all entries differ. */
|
2005-04-02 10:21:56 +00:00
|
|
|
write_c0_entryhi(UNIQUE_ENTRYHI(idx));
|
2005-04-16 22:20:36 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
tlbw_use_hazard();
|
|
|
|
}
|
|
|
|
write_c0_entryhi(oldpid);
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB;
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We will need multiple versions of update_mmu_cache(), one that just
|
|
|
|
* updates the TLB with the new pte(s), and another which also checks
|
|
|
|
* for the R4k "end of page" hardware bug and does the needy.
|
|
|
|
*/
|
|
|
|
void __update_tlb(struct vm_area_struct * vma, unsigned long address, pte_t pte)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
pgd_t *pgdp;
|
2005-02-10 12:19:59 +00:00
|
|
|
pud_t *pudp;
|
2005-04-16 22:20:36 +00:00
|
|
|
pmd_t *pmdp;
|
|
|
|
pte_t *ptep;
|
|
|
|
int idx, pid;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Handle debugger faulting in for debugee.
|
|
|
|
*/
|
|
|
|
if (current->active_mm != vma->vm_mm)
|
|
|
|
return;
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-02 10:21:56 +00:00
|
|
|
|
2013-05-13 20:56:44 +00:00
|
|
|
pid = read_c0_entryhi() & ASID_MASK;
|
2005-04-16 22:20:36 +00:00
|
|
|
address &= (PAGE_MASK << 1);
|
|
|
|
write_c0_entryhi(address | pid);
|
|
|
|
pgdp = pgd_offset(vma->vm_mm, address);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_probe();
|
2006-09-08 02:16:21 +00:00
|
|
|
tlb_probe_hazard();
|
2005-02-10 12:19:59 +00:00
|
|
|
pudp = pud_offset(pgdp, address);
|
|
|
|
pmdp = pmd_offset(pudp, address);
|
2005-04-16 22:20:36 +00:00
|
|
|
idx = read_c0_index();
|
2012-10-16 22:48:10 +00:00
|
|
|
#ifdef CONFIG_MIPS_HUGE_TLB_SUPPORT
|
2009-05-28 00:47:44 +00:00
|
|
|
/* this could be a huge page */
|
|
|
|
if (pmd_huge(*pmdp)) {
|
|
|
|
unsigned long lo;
|
|
|
|
write_c0_pagemask(PM_HUGE_MASK);
|
|
|
|
ptep = (pte_t *)pmdp;
|
2010-02-10 23:12:47 +00:00
|
|
|
lo = pte_to_entrylo(pte_val(*ptep));
|
2009-05-28 00:47:44 +00:00
|
|
|
write_c0_entrylo0(lo);
|
|
|
|
write_c0_entrylo1(lo + (HPAGE_SIZE >> 7));
|
|
|
|
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
if (idx < 0)
|
|
|
|
tlb_write_random();
|
|
|
|
else
|
|
|
|
tlb_write_indexed();
|
2012-10-16 23:01:21 +00:00
|
|
|
tlbw_use_hazard();
|
2009-05-28 00:47:44 +00:00
|
|
|
write_c0_pagemask(PM_DEFAULT_MASK);
|
|
|
|
} else
|
|
|
|
#endif
|
|
|
|
{
|
|
|
|
ptep = pte_offset_map(pmdp, address);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-09-18 23:46:32 +00:00
|
|
|
#if defined(CONFIG_64BIT_PHYS_ADDR) && defined(CONFIG_CPU_MIPS32)
|
2009-05-28 00:47:44 +00:00
|
|
|
write_c0_entrylo0(ptep->pte_high);
|
|
|
|
ptep++;
|
|
|
|
write_c0_entrylo1(ptep->pte_high);
|
2005-04-16 22:20:36 +00:00
|
|
|
#else
|
2010-02-10 23:12:47 +00:00
|
|
|
write_c0_entrylo0(pte_to_entrylo(pte_val(*ptep++)));
|
|
|
|
write_c0_entrylo1(pte_to_entrylo(pte_val(*ptep)));
|
2005-04-16 22:20:36 +00:00
|
|
|
#endif
|
2009-05-28 00:47:44 +00:00
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
if (idx < 0)
|
|
|
|
tlb_write_random();
|
|
|
|
else
|
|
|
|
tlb_write_indexed();
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
tlbw_use_hazard();
|
2007-06-06 06:52:43 +00:00
|
|
|
FLUSH_ITLB_VM(vma);
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2011-08-02 17:51:08 +00:00
|
|
|
void add_wired_entry(unsigned long entrylo0, unsigned long entrylo1,
|
|
|
|
unsigned long entryhi, unsigned long pagemask)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
unsigned long wired;
|
|
|
|
unsigned long old_pagemask;
|
|
|
|
unsigned long old_ctx;
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
ENTER_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Save old context and create impossible VPN2 value */
|
|
|
|
old_ctx = read_c0_entryhi();
|
|
|
|
old_pagemask = read_c0_pagemask();
|
|
|
|
wired = read_c0_wired();
|
|
|
|
write_c0_wired(wired + 1);
|
|
|
|
write_c0_index(wired);
|
2006-09-08 02:16:21 +00:00
|
|
|
tlbw_use_hazard(); /* What is the hazard here? */
|
2005-04-16 22:20:36 +00:00
|
|
|
write_c0_pagemask(pagemask);
|
|
|
|
write_c0_entryhi(entryhi);
|
|
|
|
write_c0_entrylo0(entrylo0);
|
|
|
|
write_c0_entrylo1(entrylo1);
|
|
|
|
mtc0_tlbw_hazard();
|
|
|
|
tlb_write_indexed();
|
|
|
|
tlbw_use_hazard();
|
|
|
|
|
|
|
|
write_c0_entryhi(old_ctx);
|
2006-09-08 02:16:21 +00:00
|
|
|
tlbw_use_hazard(); /* What is the hazard here? */
|
2005-04-16 22:20:36 +00:00
|
|
|
write_c0_pagemask(old_pagemask);
|
|
|
|
local_flush_tlb_all();
|
2006-04-05 08:45:45 +00:00
|
|
|
EXIT_CRITICAL(flags);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2012-10-18 11:54:15 +00:00
|
|
|
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
|
|
|
|
|
int __init has_transparent_hugepage(void)
|
|
|
|
{
|
|
|
|
unsigned int mask;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
ENTER_CRITICAL(flags);
|
|
|
|
write_c0_pagemask(PM_HUGE_MASK);
|
|
|
|
back_to_back_c0_hazard();
|
|
|
|
mask = read_c0_pagemask();
|
|
|
|
write_c0_pagemask(PM_DEFAULT_MASK);
|
|
|
|
|
|
|
|
EXIT_CRITICAL(flags);
|
|
|
|
|
|
|
|
return mask == PM_HUGE_MASK;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
|
|
|
|
|
MIPS: Delete __cpuinit/__CPUINIT usage from MIPS code
commit 3747069b25e419f6b51395f48127e9812abc3596 upstream.
The __cpuinit type of throwaway sections might have made sense
some time ago when RAM was more constrained, but now the savings
do not offset the cost and complications. For example, the fix in
commit 5e427ec2d0 ("x86: Fix bit corruption at CPU resume time")
is a good example of the nasty type of bugs that can be created
with improper use of the various __init prefixes.
After a discussion on LKML[1] it was decided that cpuinit should go
the way of devinit and be phased out. Once all the users are gone,
we can then finally remove the macros themselves from linux/init.h.
Note that some harmless section mismatch warnings may result, since
notify_cpu_starting() and cpu_up() are arch independent (kernel/cpu.c)
and are flagged as __cpuinit -- so if we remove the __cpuinit from
the arch specific callers, we will also get section mismatch warnings.
As an intermediate step, we intend to turn the linux/init.h cpuinit
related content into no-ops as early as possible, since that will get
rid of these warnings. In any case, they are temporary and harmless.
Here, we remove all the MIPS __cpuinit from C code and __CPUINIT
from asm files. MIPS is interesting in this respect, because there
are also uasm users hiding behind their own renamed versions of the
__cpuinit macros.
[1] https://lkml.org/lkml/2013/5/20/589
[ralf@linux-mips.org: Folded in Paul's followup fix.]
Signed-off-by: Paul Gortmaker <paul.gortmaker@windriver.com>
Cc: linux-mips@linux-mips.org
Patchwork: https://patchwork.linux-mips.org/patch/5494/
Patchwork: https://patchwork.linux-mips.org/patch/5495/
Patchwork: https://patchwork.linux-mips.org/patch/5509/
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
2013-06-18 13:38:59 +00:00
|
|
|
static int ntlb;
|
2006-04-05 08:45:45 +00:00
|
|
|
static int __init set_ntlb(char *str)
|
|
|
|
{
|
|
|
|
get_option(&str, &ntlb);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
__setup("ntlb=", set_ntlb);
|
|
|
|
|
MIPS: Delete __cpuinit/__CPUINIT usage from MIPS code
commit 3747069b25e419f6b51395f48127e9812abc3596 upstream.
The __cpuinit type of throwaway sections might have made sense
some time ago when RAM was more constrained, but now the savings
do not offset the cost and complications. For example, the fix in
commit 5e427ec2d0 ("x86: Fix bit corruption at CPU resume time")
is a good example of the nasty type of bugs that can be created
with improper use of the various __init prefixes.
After a discussion on LKML[1] it was decided that cpuinit should go
the way of devinit and be phased out. Once all the users are gone,
we can then finally remove the macros themselves from linux/init.h.
Note that some harmless section mismatch warnings may result, since
notify_cpu_starting() and cpu_up() are arch independent (kernel/cpu.c)
and are flagged as __cpuinit -- so if we remove the __cpuinit from
the arch specific callers, we will also get section mismatch warnings.
As an intermediate step, we intend to turn the linux/init.h cpuinit
related content into no-ops as early as possible, since that will get
rid of these warnings. In any case, they are temporary and harmless.
Here, we remove all the MIPS __cpuinit from C code and __CPUINIT
from asm files. MIPS is interesting in this respect, because there
are also uasm users hiding behind their own renamed versions of the
__cpuinit macros.
[1] https://lkml.org/lkml/2013/5/20/589
[ralf@linux-mips.org: Folded in Paul's followup fix.]
Signed-off-by: Paul Gortmaker <paul.gortmaker@windriver.com>
Cc: linux-mips@linux-mips.org
Patchwork: https://patchwork.linux-mips.org/patch/5494/
Patchwork: https://patchwork.linux-mips.org/patch/5495/
Patchwork: https://patchwork.linux-mips.org/patch/5509/
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
2013-06-18 13:38:59 +00:00
|
|
|
void tlb_init(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* You should never change this register:
|
|
|
|
* - On R4600 1.7 the tlbp never hits for pages smaller than
|
|
|
|
* the value in the c0_pagemask register.
|
|
|
|
* - The entire mm handling assumes the c0_pagemask register to
|
2008-02-29 00:43:47 +00:00
|
|
|
* be set to fixed-size pages.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
write_c0_pagemask(PM_DEFAULT_MASK);
|
|
|
|
write_c0_wired(0);
|
2009-01-06 23:07:20 +00:00
|
|
|
if (current_cpu_type() == CPU_R10000 ||
|
|
|
|
current_cpu_type() == CPU_R12000 ||
|
|
|
|
current_cpu_type() == CPU_R14000)
|
|
|
|
write_c0_framemask(0);
|
2010-02-10 23:12:47 +00:00
|
|
|
|
2012-09-13 21:51:46 +00:00
|
|
|
if (cpu_has_rixi) {
|
2010-02-10 23:12:47 +00:00
|
|
|
/*
|
|
|
|
* Enable the no read, no exec bits, and enable large virtual
|
|
|
|
* address.
|
|
|
|
*/
|
|
|
|
u32 pg = PG_RIE | PG_XIE;
|
|
|
|
#ifdef CONFIG_64BIT
|
|
|
|
pg |= PG_ELPA;
|
|
|
|
#endif
|
|
|
|
write_c0_pagegrain(pg);
|
|
|
|
}
|
|
|
|
|
2013-01-22 11:59:30 +00:00
|
|
|
/* From this point on the ARC firmware is dead. */
|
2005-04-16 22:20:36 +00:00
|
|
|
local_flush_tlb_all();
|
|
|
|
|
2006-03-14 14:35:27 +00:00
|
|
|
/* Did I tell you that ARC SUCKS? */
|
|
|
|
|
2006-04-05 08:45:45 +00:00
|
|
|
if (ntlb) {
|
|
|
|
if (ntlb > 1 && ntlb <= current_cpu_data.tlbsize) {
|
|
|
|
int wired = current_cpu_data.tlbsize - ntlb;
|
|
|
|
write_c0_wired(wired);
|
|
|
|
write_c0_index(wired-1);
|
2007-10-11 22:46:15 +00:00
|
|
|
printk("Restricting TLB to %d entries\n", ntlb);
|
2006-04-05 08:45:45 +00:00
|
|
|
} else
|
|
|
|
printk("Ignoring invalid argument ntlb=%d\n", ntlb);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
build_tlb_refill_handler();
|
|
|
|
}
|