forked from luck/tmp_suning_uos_patched
805918f80f
Signed-off-by: Sam Ravnborg <sam@ravnborg.org> Cc: Daniel Hellstrom <daniel@gaisler.com> Cc: Konrad Eisele <konrad@gaisler.com>
352 lines
8.2 KiB
C
352 lines
8.2 KiB
C
/*
|
|
* linux/arch/sparc/mm/leon_m.c
|
|
*
|
|
* Copyright (C) 2004 Konrad Eisele (eiselekd@web.de, konrad@gaisler.com) Gaisler Research
|
|
* Copyright (C) 2009 Daniel Hellstrom (daniel@gaisler.com) Aeroflex Gaisler AB
|
|
* Copyright (C) 2009 Konrad Eisele (konrad@gaisler.com) Aeroflex Gaisler AB
|
|
*
|
|
* do srmmu probe in software
|
|
*
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/mm.h>
|
|
#include <asm/asi.h>
|
|
#include <asm/leon.h>
|
|
#include <asm/tlbflush.h>
|
|
|
|
#include "srmmu.h"
|
|
|
|
int leon_flush_during_switch = 1;
|
|
int srmmu_swprobe_trace;
|
|
|
|
static inline unsigned long leon_get_ctable_ptr(void)
|
|
{
|
|
unsigned int retval;
|
|
|
|
__asm__ __volatile__("lda [%1] %2, %0\n\t" :
|
|
"=r" (retval) :
|
|
"r" (SRMMU_CTXTBL_PTR),
|
|
"i" (ASI_LEON_MMUREGS));
|
|
return (retval & SRMMU_CTX_PMASK) << 4;
|
|
}
|
|
|
|
|
|
unsigned long leon_swprobe(unsigned long vaddr, unsigned long *paddr)
|
|
{
|
|
|
|
unsigned int ctxtbl;
|
|
unsigned int pgd, pmd, ped;
|
|
unsigned int ptr;
|
|
unsigned int lvl, pte, paddrbase;
|
|
unsigned int ctx;
|
|
unsigned int paddr_calc;
|
|
|
|
paddrbase = 0;
|
|
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: trace on\n");
|
|
|
|
ctxtbl = leon_get_ctable_ptr();
|
|
if (!(ctxtbl)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: leon_get_ctable_ptr returned 0=>0\n");
|
|
return 0;
|
|
}
|
|
if (!_pfn_valid(PFN(ctxtbl))) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO
|
|
"swprobe: !_pfn_valid(%x)=>0\n",
|
|
PFN(ctxtbl));
|
|
return 0;
|
|
}
|
|
|
|
ctx = srmmu_get_context();
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: --- ctx (%x) ---\n", ctx);
|
|
|
|
pgd = LEON_BYPASS_LOAD_PA(ctxtbl + (ctx * 4));
|
|
|
|
if (((pgd & SRMMU_ET_MASK) == SRMMU_ET_PTE)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: pgd is entry level 3\n");
|
|
lvl = 3;
|
|
pte = pgd;
|
|
paddrbase = pgd & _SRMMU_PTE_PMASK_LEON;
|
|
goto ready;
|
|
}
|
|
if (((pgd & SRMMU_ET_MASK) != SRMMU_ET_PTD)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: pgd is invalid => 0\n");
|
|
return 0;
|
|
}
|
|
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: --- pgd (%x) ---\n", pgd);
|
|
|
|
ptr = (pgd & SRMMU_PTD_PMASK) << 4;
|
|
ptr += ((((vaddr) >> LEON_PGD_SH) & LEON_PGD_M) * 4);
|
|
if (!_pfn_valid(PFN(ptr)))
|
|
return 0;
|
|
|
|
pmd = LEON_BYPASS_LOAD_PA(ptr);
|
|
if (((pmd & SRMMU_ET_MASK) == SRMMU_ET_PTE)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: pmd is entry level 2\n");
|
|
lvl = 2;
|
|
pte = pmd;
|
|
paddrbase = pmd & _SRMMU_PTE_PMASK_LEON;
|
|
goto ready;
|
|
}
|
|
if (((pmd & SRMMU_ET_MASK) != SRMMU_ET_PTD)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: pmd is invalid => 0\n");
|
|
return 0;
|
|
}
|
|
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: --- pmd (%x) ---\n", pmd);
|
|
|
|
ptr = (pmd & SRMMU_PTD_PMASK) << 4;
|
|
ptr += (((vaddr >> LEON_PMD_SH) & LEON_PMD_M) * 4);
|
|
if (!_pfn_valid(PFN(ptr))) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: !_pfn_valid(%x)=>0\n",
|
|
PFN(ptr));
|
|
return 0;
|
|
}
|
|
|
|
ped = LEON_BYPASS_LOAD_PA(ptr);
|
|
|
|
if (((ped & SRMMU_ET_MASK) == SRMMU_ET_PTE)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: ped is entry level 1\n");
|
|
lvl = 1;
|
|
pte = ped;
|
|
paddrbase = ped & _SRMMU_PTE_PMASK_LEON;
|
|
goto ready;
|
|
}
|
|
if (((ped & SRMMU_ET_MASK) != SRMMU_ET_PTD)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: ped is invalid => 0\n");
|
|
return 0;
|
|
}
|
|
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: --- ped (%x) ---\n", ped);
|
|
|
|
ptr = (ped & SRMMU_PTD_PMASK) << 4;
|
|
ptr += (((vaddr >> LEON_PTE_SH) & LEON_PTE_M) * 4);
|
|
if (!_pfn_valid(PFN(ptr)))
|
|
return 0;
|
|
|
|
ptr = LEON_BYPASS_LOAD_PA(ptr);
|
|
if (((ptr & SRMMU_ET_MASK) == SRMMU_ET_PTE)) {
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: ptr is entry level 0\n");
|
|
lvl = 0;
|
|
pte = ptr;
|
|
paddrbase = ptr & _SRMMU_PTE_PMASK_LEON;
|
|
goto ready;
|
|
}
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: ptr is invalid => 0\n");
|
|
return 0;
|
|
|
|
ready:
|
|
switch (lvl) {
|
|
case 0:
|
|
paddr_calc =
|
|
(vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4);
|
|
break;
|
|
case 1:
|
|
paddr_calc =
|
|
(vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4);
|
|
break;
|
|
case 2:
|
|
paddr_calc =
|
|
(vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4);
|
|
break;
|
|
default:
|
|
case 3:
|
|
paddr_calc = vaddr;
|
|
break;
|
|
}
|
|
if (srmmu_swprobe_trace)
|
|
printk(KERN_INFO "swprobe: padde %x\n", paddr_calc);
|
|
if (paddr)
|
|
*paddr = paddr_calc;
|
|
return pte;
|
|
}
|
|
|
|
void leon_flush_icache_all(void)
|
|
{
|
|
__asm__ __volatile__(" flush "); /*iflush*/
|
|
}
|
|
|
|
void leon_flush_dcache_all(void)
|
|
{
|
|
__asm__ __volatile__("sta %%g0, [%%g0] %0\n\t" : :
|
|
"i"(ASI_LEON_DFLUSH) : "memory");
|
|
}
|
|
|
|
void leon_flush_pcache_all(struct vm_area_struct *vma, unsigned long page)
|
|
{
|
|
if (vma->vm_flags & VM_EXEC)
|
|
leon_flush_icache_all();
|
|
leon_flush_dcache_all();
|
|
}
|
|
|
|
void leon_flush_cache_all(void)
|
|
{
|
|
__asm__ __volatile__(" flush "); /*iflush*/
|
|
__asm__ __volatile__("sta %%g0, [%%g0] %0\n\t" : :
|
|
"i"(ASI_LEON_DFLUSH) : "memory");
|
|
}
|
|
|
|
void leon_flush_tlb_all(void)
|
|
{
|
|
leon_flush_cache_all();
|
|
__asm__ __volatile__("sta %%g0, [%0] %1\n\t" : : "r"(0x400),
|
|
"i"(ASI_LEON_MMUFLUSH) : "memory");
|
|
}
|
|
|
|
/* get all cache regs */
|
|
void leon3_getCacheRegs(struct leon3_cacheregs *regs)
|
|
{
|
|
unsigned long ccr, iccr, dccr;
|
|
|
|
if (!regs)
|
|
return;
|
|
/* Get Cache regs from "Cache ASI" address 0x0, 0x8 and 0xC */
|
|
__asm__ __volatile__("lda [%%g0] %3, %0\n\t"
|
|
"mov 0x08, %%g1\n\t"
|
|
"lda [%%g1] %3, %1\n\t"
|
|
"mov 0x0c, %%g1\n\t"
|
|
"lda [%%g1] %3, %2\n\t"
|
|
: "=r"(ccr), "=r"(iccr), "=r"(dccr)
|
|
/* output */
|
|
: "i"(ASI_LEON_CACHEREGS) /* input */
|
|
: "g1" /* clobber list */
|
|
);
|
|
regs->ccr = ccr;
|
|
regs->iccr = iccr;
|
|
regs->dccr = dccr;
|
|
}
|
|
|
|
/* Due to virtual cache we need to check cache configuration if
|
|
* it is possible to skip flushing in some cases.
|
|
*
|
|
* Leon2 and Leon3 differ in their way of telling cache information
|
|
*
|
|
*/
|
|
int __init leon_flush_needed(void)
|
|
{
|
|
int flush_needed = -1;
|
|
unsigned int ssize, sets;
|
|
char *setStr[4] =
|
|
{ "direct mapped", "2-way associative", "3-way associative",
|
|
"4-way associative"
|
|
};
|
|
/* leon 3 */
|
|
struct leon3_cacheregs cregs;
|
|
leon3_getCacheRegs(&cregs);
|
|
sets = (cregs.dccr & LEON3_XCCR_SETS_MASK) >> 24;
|
|
/* (ssize=>realsize) 0=>1k, 1=>2k, 2=>4k, 3=>8k ... */
|
|
ssize = 1 << ((cregs.dccr & LEON3_XCCR_SSIZE_MASK) >> 20);
|
|
|
|
printk(KERN_INFO "CACHE: %s cache, set size %dk\n",
|
|
sets > 3 ? "unknown" : setStr[sets], ssize);
|
|
if ((ssize <= (PAGE_SIZE / 1024)) && (sets == 0)) {
|
|
/* Set Size <= Page size ==>
|
|
flush on every context switch not needed. */
|
|
flush_needed = 0;
|
|
printk(KERN_INFO "CACHE: not flushing on every context switch\n");
|
|
}
|
|
return flush_needed;
|
|
}
|
|
|
|
void leon_switch_mm(void)
|
|
{
|
|
flush_tlb_mm((void *)0);
|
|
if (leon_flush_during_switch)
|
|
leon_flush_cache_all();
|
|
}
|
|
|
|
static void leon_flush_cache_mm(struct mm_struct *mm)
|
|
{
|
|
leon_flush_cache_all();
|
|
}
|
|
|
|
static void leon_flush_cache_page(struct vm_area_struct *vma, unsigned long page)
|
|
{
|
|
leon_flush_pcache_all(vma, page);
|
|
}
|
|
|
|
static void leon_flush_cache_range(struct vm_area_struct *vma,
|
|
unsigned long start,
|
|
unsigned long end)
|
|
{
|
|
leon_flush_cache_all();
|
|
}
|
|
|
|
static void leon_flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
leon_flush_tlb_all();
|
|
}
|
|
|
|
static void leon_flush_tlb_page(struct vm_area_struct *vma,
|
|
unsigned long page)
|
|
{
|
|
leon_flush_tlb_all();
|
|
}
|
|
|
|
static void leon_flush_tlb_range(struct vm_area_struct *vma,
|
|
unsigned long start,
|
|
unsigned long end)
|
|
{
|
|
leon_flush_tlb_all();
|
|
}
|
|
|
|
static void leon_flush_page_to_ram(unsigned long page)
|
|
{
|
|
leon_flush_cache_all();
|
|
}
|
|
|
|
static void leon_flush_sig_insns(struct mm_struct *mm, unsigned long page)
|
|
{
|
|
leon_flush_cache_all();
|
|
}
|
|
|
|
static void leon_flush_page_for_dma(unsigned long page)
|
|
{
|
|
leon_flush_dcache_all();
|
|
}
|
|
|
|
void __init poke_leonsparc(void)
|
|
{
|
|
}
|
|
|
|
static const struct sparc32_cachetlb_ops leon_ops = {
|
|
.cache_all = leon_flush_cache_all,
|
|
.cache_mm = leon_flush_cache_mm,
|
|
.cache_page = leon_flush_cache_page,
|
|
.cache_range = leon_flush_cache_range,
|
|
.tlb_all = leon_flush_tlb_all,
|
|
.tlb_mm = leon_flush_tlb_mm,
|
|
.tlb_page = leon_flush_tlb_page,
|
|
.tlb_range = leon_flush_tlb_range,
|
|
.page_to_ram = leon_flush_page_to_ram,
|
|
.sig_insns = leon_flush_sig_insns,
|
|
.page_for_dma = leon_flush_page_for_dma,
|
|
};
|
|
|
|
void __init init_leon(void)
|
|
{
|
|
srmmu_name = "LEON";
|
|
sparc32_cachetlb_ops = &leon_ops;
|
|
poke_srmmu = poke_leonsparc;
|
|
|
|
leon_flush_during_switch = leon_flush_needed();
|
|
}
|