From: Haavard Skinnemoen Date: Sun, 1 Oct 2006 06:29:14 +0000 (-0700) Subject: [PATCH] Generic ioremap_page_range: flush_cache_vmap X-Git-Tag: v2.6.19-rc1~520 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=db71daabad0821996483dfe309c4bc81d6755a70;p=platform%2Fupstream%2Fkernel-adaptation-pc.git [PATCH] Generic ioremap_page_range: flush_cache_vmap The existing implementation of ioremap_page_range(), which was taken from i386, does this: flush_cache_all(); /* modify page tables */ flush_tlb_all(); I think this is a bit defensive, so this patch changes the generic implementation to do: /* modify page tables */ flush_cache_vmap(start, end); instead, which is similar to what vmalloc() does. This should still be correct because we never modify existing PTEs. According to James Bottomley: The problem the flush_tlb_all() is trying to solve is to avoid stale tlb entries in the ioremap area. We're just being conservative by flushing on both map and unmap. Technically what vmalloc/vfree does (only flush the tlb on unmap) is just fine because it means that the only tlb entries in the remap area must belong to in-use mappings. Signed-off-by: Haavard Skinnemoen Cc: Richard Henderson Cc: Ivan Kokshaysky Cc: Russell King Cc: Mikael Starvik Cc: Andi Kleen Cc: Cc: Ralf Baechle Cc: Kyle McMartin Cc: Martin Schwidefsky Cc: Paul Mundt Signed-off-by: Andrew Morton Signed-off-by: Linus Torvalds --- diff --git a/lib/ioremap.c b/lib/ioremap.c index 29c810e..99fa277 100644 --- a/lib/ioremap.c +++ b/lib/ioremap.c @@ -76,8 +76,6 @@ int ioremap_page_range(unsigned long addr, BUG_ON(addr >= end); - flush_cache_all(); - start = addr; phys_addr -= addr; pgd = pgd_offset_k(addr); @@ -88,7 +86,7 @@ int ioremap_page_range(unsigned long addr, break; } while (pgd++, addr = next, addr != end); - flush_tlb_all(); + flush_cache_vmap(start, end); return err; }