/* SPDX-License-Identifier: GPL-2.0-only */ /* * TLB flush support for Hexagon * * Copyright (c) 2010-2011, The Linux Foundation. All rights reserved. */ #ifndef _ASM_TLBFLUSH_H #define _ASM_TLBFLUSH_H #include <linux/mm.h> #include <asm/processor.h> /* * TLB flushing -- in "SMP", these routines get defined to be the * ones from smp.c, else they are some local flavors. */ /* * These functions are commonly macros, but in the interests of * VM vs. native implementation and code size, we simply declare * the function prototypes here. */ extern void tlb_flush_all(void); extern void flush_tlb_mm(struct mm_struct *mm); extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr); extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); extern void flush_tlb_one(unsigned long); /* * "This is called in munmap when we have freed up some page-table pages. * We don't need to do anything here..." * * The VM kernel doesn't walk page tables, and they are passed to the VMM * by logical address. There doesn't seem to be any possibility that they * could be referenced by the VM kernel based on a stale mapping, since * they would only be located by consulting the mm structure, and they * will have been purged from that structure by the munmap. Seems like * a noop on HVM as well. */ #define flush_tlb_pgtables(mm, start, end) #endif