|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | // Copyright (C) 2005-2017 Andes Technology Corporation | 
|  |  | 
|  | #include <linux/module.h> | 
|  | #include <linux/sched.h> | 
|  | #include <linux/mm.h> | 
|  | #include <asm/nds32.h> | 
|  | #include <asm/pgtable.h> | 
|  | #include <asm/tlbflush.h> | 
|  | #include <asm/cacheflush.h> | 
|  | #include <asm/l2_cache.h> | 
|  | #include <nds32_intrinsic.h> | 
|  |  | 
|  | #include <asm/cache_info.h> | 
|  | extern struct cache_info L1_cache_info[2]; | 
|  |  | 
|  | int va_kernel_present(unsigned long addr) | 
|  | { | 
|  | pmd_t *pmd; | 
|  | pte_t *ptep, pte; | 
|  |  | 
|  | pmd = pmd_offset(pgd_offset_k(addr), addr); | 
|  | if (!pmd_none(*pmd)) { | 
|  | ptep = pte_offset_map(pmd, addr); | 
|  | pte = *ptep; | 
|  | if (pte_present(pte)) | 
|  | return pte; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | pte_t va_present(struct mm_struct * mm, unsigned long addr) | 
|  | { | 
|  | pgd_t *pgd; | 
|  | pud_t *pud; | 
|  | pmd_t *pmd; | 
|  | pte_t *ptep, pte; | 
|  |  | 
|  | pgd = pgd_offset(mm, addr); | 
|  | if (!pgd_none(*pgd)) { | 
|  | pud = pud_offset(pgd, addr); | 
|  | if (!pud_none(*pud)) { | 
|  | pmd = pmd_offset(pud, addr); | 
|  | if (!pmd_none(*pmd)) { | 
|  | ptep = pte_offset_map(pmd, addr); | 
|  | pte = *ptep; | 
|  | if (pte_present(pte)) | 
|  | return pte; | 
|  | } | 
|  | } | 
|  | } | 
|  | return 0; | 
|  |  | 
|  | } | 
|  |  | 
|  | int va_readable(struct pt_regs *regs, unsigned long addr) | 
|  | { | 
|  | struct mm_struct *mm = current->mm; | 
|  | pte_t pte; | 
|  | int ret = 0; | 
|  |  | 
|  | if (user_mode(regs)) { | 
|  | /* user mode */ | 
|  | pte = va_present(mm, addr); | 
|  | if (!pte && pte_read(pte)) | 
|  | ret = 1; | 
|  | } else { | 
|  | /* superuser mode is always readable, so we can only | 
|  | * check it is present or not*/ | 
|  | return (! !va_kernel_present(addr)); | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | int va_writable(struct pt_regs *regs, unsigned long addr) | 
|  | { | 
|  | struct mm_struct *mm = current->mm; | 
|  | pte_t pte; | 
|  | int ret = 0; | 
|  |  | 
|  | if (user_mode(regs)) { | 
|  | /* user mode */ | 
|  | pte = va_present(mm, addr); | 
|  | if (!pte && pte_write(pte)) | 
|  | ret = 1; | 
|  | } else { | 
|  | /* superuser mode */ | 
|  | pte = va_kernel_present(addr); | 
|  | if (!pte && pte_kernel_write(pte)) | 
|  | ret = 1; | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * All | 
|  | */ | 
|  | void cpu_icache_inval_all(void) | 
|  | { | 
|  | unsigned long end, line_size; | 
|  |  | 
|  | line_size = L1_cache_info[ICACHE].line_size; | 
|  | end = | 
|  | line_size * L1_cache_info[ICACHE].ways * L1_cache_info[ICACHE].sets; | 
|  |  | 
|  | do { | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_IX_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_IX_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_IX_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_IX_INVAL"::"r" (end)); | 
|  | } while (end > 0); | 
|  | __nds32__isb(); | 
|  | } | 
|  |  | 
|  | void cpu_dcache_inval_all(void) | 
|  | { | 
|  | __nds32__cctl_l1d_invalall(); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_CACHE_L2 | 
|  | void dcache_wb_all_level(void) | 
|  | { | 
|  | unsigned long flags, cmd; | 
|  | local_irq_save(flags); | 
|  | __nds32__cctl_l1d_wball_alvl(); | 
|  | /* Section 1: Ensure the section 2 & 3 program code execution after */ | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  |  | 
|  | /* Section 2: Confirm the writeback all level is done in CPU and L2C */ | 
|  | cmd = CCTL_CMD_L2_SYNC; | 
|  | L2_CMD_RDY(); | 
|  | L2C_W_REG(L2_CCTL_CMD_OFF, cmd); | 
|  | L2_CMD_RDY(); | 
|  |  | 
|  | /* Section 3: Writeback whole L2 cache */ | 
|  | cmd = CCTL_ALL_CMD | CCTL_CMD_L2_IX_WB; | 
|  | L2_CMD_RDY(); | 
|  | L2C_W_REG(L2_CCTL_CMD_OFF, cmd); | 
|  | L2_CMD_RDY(); | 
|  | __nds32__msync_all(); | 
|  | local_irq_restore(flags); | 
|  | } | 
|  | EXPORT_SYMBOL(dcache_wb_all_level); | 
|  | #endif | 
|  |  | 
|  | void cpu_dcache_wb_all(void) | 
|  | { | 
|  | __nds32__cctl_l1d_wball_one_lvl(); | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  | } | 
|  |  | 
|  | void cpu_dcache_wbinval_all(void) | 
|  | { | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | unsigned long flags; | 
|  | local_irq_save(flags); | 
|  | #endif | 
|  | cpu_dcache_wb_all(); | 
|  | cpu_dcache_inval_all(); | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | local_irq_restore(flags); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Page | 
|  | */ | 
|  | void cpu_icache_inval_page(unsigned long start) | 
|  | { | 
|  | unsigned long line_size, end; | 
|  |  | 
|  | line_size = L1_cache_info[ICACHE].line_size; | 
|  | end = start + PAGE_SIZE; | 
|  |  | 
|  | do { | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_VA_INVAL"::"r" (end)); | 
|  | } while (end != start); | 
|  | __nds32__isb(); | 
|  | } | 
|  |  | 
|  | void cpu_dcache_inval_page(unsigned long start) | 
|  | { | 
|  | unsigned long line_size, end; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | end = start + PAGE_SIZE; | 
|  |  | 
|  | do { | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | } while (end != start); | 
|  | } | 
|  |  | 
|  | void cpu_dcache_wb_page(unsigned long start) | 
|  | { | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | unsigned long line_size, end; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | end = start + PAGE_SIZE; | 
|  |  | 
|  | do { | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | end -= line_size; | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | } while (end != start); | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | void cpu_dcache_wbinval_page(unsigned long start) | 
|  | { | 
|  | unsigned long line_size, end; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | end = start + PAGE_SIZE; | 
|  |  | 
|  | do { | 
|  | end -= line_size; | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | #endif | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | #endif | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | #endif | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | end -= line_size; | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (end)); | 
|  | #endif | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (end)); | 
|  | } while (end != start); | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  | } | 
|  |  | 
|  | void cpu_cache_wbinval_page(unsigned long page, int flushi) | 
|  | { | 
|  | cpu_dcache_wbinval_page(page); | 
|  | if (flushi) | 
|  | cpu_icache_inval_page(page); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Range | 
|  | */ | 
|  | void cpu_icache_inval_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  |  | 
|  | line_size = L1_cache_info[ICACHE].line_size; | 
|  |  | 
|  | while (end > start) { | 
|  | __asm__ volatile ("\n\tcctl %0, L1I_VA_INVAL"::"r" (start)); | 
|  | start += line_size; | 
|  | } | 
|  | __nds32__isb(); | 
|  | } | 
|  |  | 
|  | void cpu_dcache_inval_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  |  | 
|  | while (end > start) { | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (start)); | 
|  | start += line_size; | 
|  | } | 
|  | } | 
|  |  | 
|  | void cpu_dcache_wb_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | unsigned long line_size; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  |  | 
|  | while (end > start) { | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (start)); | 
|  | start += line_size; | 
|  | } | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | void cpu_dcache_wbinval_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  |  | 
|  | while (end > start) { | 
|  | #ifndef CONFIG_CPU_DCACHE_WRITETHROUGH | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_WB"::"r" (start)); | 
|  | #endif | 
|  | __asm__ volatile ("\n\tcctl %0, L1D_VA_INVAL"::"r" (start)); | 
|  | start += line_size; | 
|  | } | 
|  | __nds32__cctlidx_read(NDS32_CCTL_L1D_IX_RWD,0); | 
|  | } | 
|  |  | 
|  | void cpu_cache_wbinval_range(unsigned long start, unsigned long end, int flushi) | 
|  | { | 
|  | unsigned long line_size, align_start, align_end; | 
|  |  | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | align_start = start & ~(line_size - 1); | 
|  | align_end = (end + line_size - 1) & ~(line_size - 1); | 
|  | cpu_dcache_wbinval_range(align_start, align_end); | 
|  |  | 
|  | if (flushi) { | 
|  | line_size = L1_cache_info[ICACHE].line_size; | 
|  | align_start = start & ~(line_size - 1); | 
|  | align_end = (end + line_size - 1) & ~(line_size - 1); | 
|  | cpu_icache_inval_range(align_start, align_end); | 
|  | } | 
|  | } | 
|  |  | 
|  | void cpu_cache_wbinval_range_check(struct vm_area_struct *vma, | 
|  | unsigned long start, unsigned long end, | 
|  | bool flushi, bool wbd) | 
|  | { | 
|  | unsigned long line_size, t_start, t_end; | 
|  |  | 
|  | if (!flushi && !wbd) | 
|  | return; | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | start = start & ~(line_size - 1); | 
|  | end = (end + line_size - 1) & ~(line_size - 1); | 
|  |  | 
|  | if ((end - start) > (8 * PAGE_SIZE)) { | 
|  | if (wbd) | 
|  | cpu_dcache_wbinval_all(); | 
|  | if (flushi) | 
|  | cpu_icache_inval_all(); | 
|  | return; | 
|  | } | 
|  |  | 
|  | t_start = (start + PAGE_SIZE) & PAGE_MASK; | 
|  | t_end = ((end - 1) & PAGE_MASK); | 
|  |  | 
|  | if ((start & PAGE_MASK) == t_end) { | 
|  | if (va_present(vma->vm_mm, start)) { | 
|  | if (wbd) | 
|  | cpu_dcache_wbinval_range(start, end); | 
|  | if (flushi) | 
|  | cpu_icache_inval_range(start, end); | 
|  | } | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (va_present(vma->vm_mm, start)) { | 
|  | if (wbd) | 
|  | cpu_dcache_wbinval_range(start, t_start); | 
|  | if (flushi) | 
|  | cpu_icache_inval_range(start, t_start); | 
|  | } | 
|  |  | 
|  | if (va_present(vma->vm_mm, end - 1)) { | 
|  | if (wbd) | 
|  | cpu_dcache_wbinval_range(t_end, end); | 
|  | if (flushi) | 
|  | cpu_icache_inval_range(t_end, end); | 
|  | } | 
|  |  | 
|  | while (t_start < t_end) { | 
|  | if (va_present(vma->vm_mm, t_start)) { | 
|  | if (wbd) | 
|  | cpu_dcache_wbinval_page(t_start); | 
|  | if (flushi) | 
|  | cpu_icache_inval_page(t_start); | 
|  | } | 
|  | t_start += PAGE_SIZE; | 
|  | } | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_CACHE_L2 | 
|  | static inline void cpu_l2cache_op(unsigned long start, unsigned long end, unsigned long op) | 
|  | { | 
|  | if (atl2c_base) { | 
|  | unsigned long p_start = __pa(start); | 
|  | unsigned long p_end = __pa(end); | 
|  | unsigned long cmd; | 
|  | unsigned long line_size; | 
|  | /* TODO Can Use PAGE Mode to optimize if range large than PAGE_SIZE */ | 
|  | line_size = L2_CACHE_LINE_SIZE(); | 
|  | p_start = p_start & (~(line_size - 1)); | 
|  | p_end = (p_end + line_size - 1) & (~(line_size - 1)); | 
|  | cmd = | 
|  | (p_start & ~(line_size - 1)) | op | | 
|  | CCTL_SINGLE_CMD; | 
|  | do { | 
|  | L2_CMD_RDY(); | 
|  | L2C_W_REG(L2_CCTL_CMD_OFF, cmd); | 
|  | cmd += line_size; | 
|  | p_start += line_size; | 
|  | } while (p_end > p_start); | 
|  | cmd = CCTL_CMD_L2_SYNC; | 
|  | L2_CMD_RDY(); | 
|  | L2C_W_REG(L2_CCTL_CMD_OFF, cmd); | 
|  | L2_CMD_RDY(); | 
|  | } | 
|  | } | 
|  | #else | 
|  | #define cpu_l2cache_op(start,end,op) do { } while (0) | 
|  | #endif | 
|  | /* | 
|  | * DMA | 
|  | */ | 
|  | void cpu_dma_wb_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  | unsigned long flags; | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | start = start & (~(line_size - 1)); | 
|  | end = (end + line_size - 1) & (~(line_size - 1)); | 
|  | if (unlikely(start == end)) | 
|  | return; | 
|  |  | 
|  | local_irq_save(flags); | 
|  | cpu_dcache_wb_range(start, end); | 
|  | cpu_l2cache_op(start, end, CCTL_CMD_L2_PA_WB); | 
|  | __nds32__msync_all(); | 
|  | local_irq_restore(flags); | 
|  | } | 
|  |  | 
|  | void cpu_dma_inval_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  | unsigned long old_start = start; | 
|  | unsigned long old_end = end; | 
|  | unsigned long flags; | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | start = start & (~(line_size - 1)); | 
|  | end = (end + line_size - 1) & (~(line_size - 1)); | 
|  | if (unlikely(start == end)) | 
|  | return; | 
|  | local_irq_save(flags); | 
|  | if (start != old_start) { | 
|  | cpu_dcache_wbinval_range(start, start + line_size); | 
|  | cpu_l2cache_op(start, start + line_size, CCTL_CMD_L2_PA_WBINVAL); | 
|  | } | 
|  | if (end != old_end) { | 
|  | cpu_dcache_wbinval_range(end - line_size, end); | 
|  | cpu_l2cache_op(end - line_size, end, CCTL_CMD_L2_PA_WBINVAL); | 
|  | } | 
|  | cpu_dcache_inval_range(start, end); | 
|  | cpu_l2cache_op(start, end, CCTL_CMD_L2_PA_INVAL); | 
|  | __nds32__msync_all(); | 
|  | local_irq_restore(flags); | 
|  |  | 
|  | } | 
|  |  | 
|  | void cpu_dma_wbinval_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long line_size; | 
|  | unsigned long flags; | 
|  | line_size = L1_cache_info[DCACHE].line_size; | 
|  | start = start & (~(line_size - 1)); | 
|  | end = (end + line_size - 1) & (~(line_size - 1)); | 
|  | if (unlikely(start == end)) | 
|  | return; | 
|  |  | 
|  | local_irq_save(flags); | 
|  | cpu_dcache_wbinval_range(start, end); | 
|  | cpu_l2cache_op(start, end, CCTL_CMD_L2_PA_WBINVAL); | 
|  | __nds32__msync_all(); | 
|  | local_irq_restore(flags); | 
|  | } | 
|  |  | 
|  | void cpu_proc_init(void) | 
|  | { | 
|  | } | 
|  |  | 
|  | void cpu_proc_fin(void) | 
|  | { | 
|  | } | 
|  |  | 
|  | void cpu_do_idle(void) | 
|  | { | 
|  | __nds32__standby_no_wake_grant(); | 
|  | } | 
|  |  | 
|  | void cpu_reset(unsigned long reset) | 
|  | { | 
|  | u32 tmp; | 
|  | GIE_DISABLE(); | 
|  | tmp = __nds32__mfsr(NDS32_SR_CACHE_CTL); | 
|  | tmp &= ~(CACHE_CTL_mskIC_EN | CACHE_CTL_mskDC_EN); | 
|  | __nds32__mtsr_isb(tmp, NDS32_SR_CACHE_CTL); | 
|  | cpu_dcache_wbinval_all(); | 
|  | cpu_icache_inval_all(); | 
|  |  | 
|  | __asm__ __volatile__("jr.toff %0\n\t"::"r"(reset)); | 
|  | } | 
|  |  | 
|  | void cpu_switch_mm(struct mm_struct *mm) | 
|  | { | 
|  | unsigned long cid; | 
|  | cid = __nds32__mfsr(NDS32_SR_TLB_MISC); | 
|  | cid = (cid & ~TLB_MISC_mskCID) | mm->context.id; | 
|  | __nds32__mtsr_dsb(cid, NDS32_SR_TLB_MISC); | 
|  | __nds32__mtsr_isb(__pa(mm->pgd), NDS32_SR_L1_PPTB); | 
|  | } |