| Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 1 | #ifndef _ASM_X86_CACHEFLUSH_H | 
 | 2 | #define _ASM_X86_CACHEFLUSH_H | 
 | 3 |  | 
 | 4 | /* Keep includes the same across arches.  */ | 
 | 5 | #include <linux/mm.h> | 
 | 6 |  | 
 | 7 | /* Caches aren't brain-dead on the intel. */ | 
 | 8 | #define flush_cache_all()			do { } while (0) | 
 | 9 | #define flush_cache_mm(mm)			do { } while (0) | 
 | 10 | #define flush_cache_dup_mm(mm)			do { } while (0) | 
 | 11 | #define flush_cache_range(vma, start, end)	do { } while (0) | 
 | 12 | #define flush_cache_page(vma, vmaddr, pfn)	do { } while (0) | 
 | 13 | #define flush_dcache_page(page)			do { } while (0) | 
 | 14 | #define flush_dcache_mmap_lock(mapping)		do { } while (0) | 
 | 15 | #define flush_dcache_mmap_unlock(mapping)	do { } while (0) | 
 | 16 | #define flush_icache_range(start, end)		do { } while (0) | 
 | 17 | #define flush_icache_page(vma,pg)		do { } while (0) | 
 | 18 | #define flush_icache_user_range(vma,pg,adr,len)	do { } while (0) | 
 | 19 | #define flush_cache_vmap(start, end)		do { } while (0) | 
 | 20 | #define flush_cache_vunmap(start, end)		do { } while (0) | 
 | 21 |  | 
 | 22 | #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ | 
 | 23 | 	memcpy(dst, src, len) | 
 | 24 | #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ | 
 | 25 | 	memcpy(dst, src, len) | 
 | 26 |  | 
| Arjan van de Ven | e1271f6 | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 27 | int __deprecated_for_modules change_page_attr(struct page *page, int numpages, | 
 | 28 | 								pgprot_t prot); | 
| Arjan van de Ven | 75cbade | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 29 |  | 
 | 30 | int set_pages_uc(struct page *page, int numpages); | 
 | 31 | int set_pages_wb(struct page *page, int numpages); | 
 | 32 | int set_pages_x(struct page *page, int numpages); | 
 | 33 | int set_pages_nx(struct page *page, int numpages); | 
 | 34 | int set_pages_ro(struct page *page, int numpages); | 
 | 35 | int set_pages_rw(struct page *page, int numpages); | 
 | 36 |  | 
 | 37 | int set_memory_uc(unsigned long addr, int numpages); | 
 | 38 | int set_memory_wb(unsigned long addr, int numpages); | 
 | 39 | int set_memory_x(unsigned long addr, int numpages); | 
 | 40 | int set_memory_nx(unsigned long addr, int numpages); | 
 | 41 | int set_memory_ro(unsigned long addr, int numpages); | 
 | 42 | int set_memory_rw(unsigned long addr, int numpages); | 
| Ingo Molnar | f62d0f0 | 2008-01-30 13:34:07 +0100 | [diff] [blame] | 43 | int set_memory_np(unsigned long addr, int numpages); | 
| Arjan van de Ven | 75cbade | 2008-01-30 13:34:06 +0100 | [diff] [blame] | 44 |  | 
| Keshavamurthy, Anil S | a9c55b3 | 2007-10-21 16:41:47 -0700 | [diff] [blame] | 45 | void clflush_cache_range(void *addr, int size); | 
| Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 46 |  | 
| Thomas Gleixner | b2bba72 | 2007-10-15 23:28:20 +0200 | [diff] [blame] | 47 | #ifdef CONFIG_DEBUG_RODATA | 
 | 48 | void mark_rodata_ro(void); | 
 | 49 | #endif | 
 | 50 |  | 
| Thomas Gleixner | 96a388d | 2007-10-11 11:20:03 +0200 | [diff] [blame] | 51 | #endif |