File manager - Edit - /home/newsbmcs.com/public_html/static/img/logo/cacheflush.h.tar
Back
usr/src/linux-headers-5.15.0-133/arch/parisc/include/asm/cacheflush.h 0000644 00000007275 15030064415 0020634 0 ustar 00 /* SPDX-License-Identifier: GPL-2.0 */ #ifndef _PARISC_CACHEFLUSH_H #define _PARISC_CACHEFLUSH_H #include <linux/mm.h> #include <linux/uaccess.h> #include <asm/tlbflush.h> /* The usual comment is "Caches aren't brain-dead on the <architecture>". * Unfortunately, that doesn't apply to PA-RISC. */ /* Internal implementation */ void flush_data_cache_local(void *); /* flushes local data-cache only */ void flush_instruction_cache_local(void *); /* flushes local code-cache only */ #ifdef CONFIG_SMP void flush_data_cache(void); /* flushes data-cache only (all processors) */ void flush_instruction_cache(void); /* flushes i-cache only (all processors) */ #else #define flush_data_cache() flush_data_cache_local(NULL) #define flush_instruction_cache() flush_instruction_cache_local(NULL) #endif #define flush_cache_dup_mm(mm) flush_cache_mm(mm) void flush_user_icache_range_asm(unsigned long, unsigned long); void flush_kernel_icache_range_asm(unsigned long, unsigned long); void flush_user_dcache_range_asm(unsigned long, unsigned long); void flush_kernel_dcache_range_asm(unsigned long, unsigned long); void purge_kernel_dcache_range_asm(unsigned long, unsigned long); void flush_kernel_dcache_page_asm(void *); void flush_kernel_icache_page(void *); /* Cache flush operations */ void flush_cache_all_local(void); void flush_cache_all(void); void flush_cache_mm(struct mm_struct *mm); void flush_kernel_dcache_page_addr(void *addr); #define flush_kernel_dcache_range(start,size) \ flush_kernel_dcache_range_asm((start), (start)+(size)); #define ARCH_IMPLEMENTS_FLUSH_KERNEL_VMAP_RANGE 1 void flush_kernel_vmap_range(void *vaddr, int size); void invalidate_kernel_vmap_range(void *vaddr, int size); #define flush_cache_vmap(start, end) flush_cache_all() #define flush_cache_vunmap(start, end) flush_cache_all() #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1 extern void flush_dcache_page(struct page *page); #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) #define flush_dcache_mmap_lock_irqsave(mapping, flags) \ xa_lock_irqsave(&mapping->i_pages, flags) #define flush_dcache_mmap_unlock_irqrestore(mapping, flags) \ xa_unlock_irqrestore(&mapping->i_pages, flags) #define flush_icache_page(vma,page) do { \ flush_kernel_dcache_page_addr(page_address(page)); \ flush_kernel_icache_page(page_address(page)); \ } while (0) #define flush_icache_range(s,e) do { \ flush_kernel_dcache_range_asm(s,e); \ flush_kernel_icache_range_asm(s,e); \ } while (0) #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ do { \ flush_cache_page(vma, vaddr, page_to_pfn(page)); \ memcpy(dst, src, len); \ flush_kernel_dcache_range_asm((unsigned long)dst, (unsigned long)dst + len); \ } while (0) #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ do { \ flush_cache_page(vma, vaddr, page_to_pfn(page)); \ memcpy(dst, src, len); \ } while (0) void flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn); void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); /* defined in pacache.S exported in cache.c used by flush_anon_page */ void flush_dcache_page_asm(unsigned long phys_addr, unsigned long vaddr); #define ARCH_HAS_FLUSH_ANON_PAGE static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) { if (PageAnon(page)) { flush_tlb_page(vma, vmaddr); preempt_disable(); flush_dcache_page_asm(page_to_phys(page), vmaddr); preempt_enable(); } } #define ARCH_HAS_FLUSH_ON_KUNMAP static inline void kunmap_flush_on_unmap(void *addr) { flush_kernel_dcache_page_addr(addr); } #endif /* _PARISC_CACHEFLUSH_H */ usr/src/linux-headers-5.15.0-133/arch/arc/include/asm/cacheflush.h 0000644 00000007360 15030123454 0020112 0 ustar 00 /* SPDX-License-Identifier: GPL-2.0-only */ /* * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com) * * vineetg: May 2011: for Non-aliasing VIPT D-cache following can be NOPs * -flush_cache_dup_mm (fork) * -likewise for flush_cache_mm (exit/execve) * -likewise for flush_cache_{range,page} (munmap, exit, COW-break) * * vineetg: April 2008 * -Added a critical CacheLine flush to copy_to_user_page( ) which * was causing gdbserver to not setup breakpoints consistently */ #ifndef _ASM_CACHEFLUSH_H #define _ASM_CACHEFLUSH_H #include <linux/mm.h> #include <asm/shmparam.h> /* * Semantically we need this because icache doesn't snoop dcache/dma. * However ARC Cache flush requires paddr as well as vaddr, latter not available * in the flush_icache_page() API. So we no-op it but do the equivalent work * in update_mmu_cache() */ #define flush_icache_page(vma, page) void flush_cache_all(void); void flush_icache_range(unsigned long kstart, unsigned long kend); void __sync_icache_dcache(phys_addr_t paddr, unsigned long vaddr, int len); void __inv_icache_page(phys_addr_t paddr, unsigned long vaddr); void __flush_dcache_page(phys_addr_t paddr, unsigned long vaddr); #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1 void flush_dcache_page(struct page *page); void dma_cache_wback_inv(phys_addr_t start, unsigned long sz); void dma_cache_inv(phys_addr_t start, unsigned long sz); void dma_cache_wback(phys_addr_t start, unsigned long sz); #define flush_dcache_mmap_lock(mapping) do { } while (0) #define flush_dcache_mmap_unlock(mapping) do { } while (0) /* TBD: optimize this */ #define flush_cache_vmap(start, end) flush_cache_all() #define flush_cache_vunmap(start, end) flush_cache_all() #define flush_cache_dup_mm(mm) /* called on fork (VIVT only) */ #ifndef CONFIG_ARC_CACHE_VIPT_ALIASING #define flush_cache_mm(mm) /* called on munmap/exit */ #define flush_cache_range(mm, u_vstart, u_vend) #define flush_cache_page(vma, u_vaddr, pfn) /* PF handling/COW-break */ #else /* VIPT aliasing dcache */ /* To clear out stale userspace mappings */ void flush_cache_mm(struct mm_struct *mm); void flush_cache_range(struct vm_area_struct *vma, unsigned long start,unsigned long end); void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long page); /* * To make sure that userspace mapping is flushed to memory before * get_user_pages() uses a kernel mapping to access the page */ #define ARCH_HAS_FLUSH_ANON_PAGE void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long u_vaddr); #endif /* CONFIG_ARC_CACHE_VIPT_ALIASING */ /* * A new pagecache page has PG_arch_1 clear - thus dcache dirty by default * This works around some PIO based drivers which don't call flush_dcache_page * to record that they dirtied the dcache */ #define PG_dc_clean PG_arch_1 #define CACHE_COLORS_NUM 4 #define CACHE_COLORS_MSK (CACHE_COLORS_NUM - 1) #define CACHE_COLOR(addr) (((unsigned long)(addr) >> (PAGE_SHIFT)) & CACHE_COLORS_MSK) /* * Simple wrapper over config option * Bootup code ensures that hardware matches kernel configuration */ static inline int cache_is_vipt_aliasing(void) { return IS_ENABLED(CONFIG_ARC_CACHE_VIPT_ALIASING); } /* * checks if two addresses (after page aligning) index into same cache set */ #define addr_not_cache_congruent(addr1, addr2) \ ({ \ cache_is_vipt_aliasing() ? \ (CACHE_COLOR(addr1) != CACHE_COLOR(addr2)) : 0; \ }) #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ do { \ memcpy(dst, src, len); \ if (vma->vm_flags & VM_EXEC) \ __sync_icache_dcache((unsigned long)(dst), vaddr, len); \ } while (0) #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ memcpy(dst, src, len); \ #endif
| ver. 1.4 |
Github
|
.
| PHP 8.2.28 | Generation time: 0.02 |
proxy
|
phpinfo
|
Settings