aboutsummaryrefslogtreecommitdiff
path: root/arch
diff options
context:
space:
mode:
authorThomas Gleixner2008-01-30 13:34:05 +0100
committerIngo Molnar2008-01-30 13:34:05 +0100
commit91eebf40b3cb5abd76e813e17dbc320ff2ea3295 (patch)
treea94d05d8d87a6b03ad26d09a8ae08087d90dfee6 /arch
parent1aaf74e919be54be0023c3124923fb537c7fb772 (diff)
x86: style cleanup of ioremap code
Fix the coding style before going further. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r--arch/x86/mm/ioremap_32.c97
-rw-r--r--arch/x86/mm/ioremap_64.c48
2 files changed, 70 insertions, 75 deletions
diff --git a/arch/x86/mm/ioremap_32.c b/arch/x86/mm/ioremap_32.c
index 18757f058bda..30ff2586db1e 100644
--- a/arch/x86/mm/ioremap_32.c
+++ b/arch/x86/mm/ioremap_32.c
@@ -1,6 +1,4 @@
/*
- * arch/i386/mm/ioremap.c
- *
* Re-map IO memory to kernel address space so that we can access it.
* This is needed for high PCI addresses that aren't mapped in the
* 640k-1MB IO memory area on PC's
@@ -22,10 +20,6 @@
#define ISA_END_ADDRESS 0x100000
/*
- * Generic mapping function (not visible outside):
- */
-
-/*
* Remap an arbitrary physical address space into the kernel virtual
* address space. Needed when the kernel wants to access high addresses
* directly.
@@ -34,10 +28,11 @@
* have to convert them into an offset in a page-aligned mapping, but the
* caller shouldn't need to know that small detail.
*/
-void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned long flags)
+void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
+ unsigned long flags)
{
- void __iomem * addr;
- struct vm_struct * area;
+ void __iomem *addr;
+ struct vm_struct *area;
unsigned long offset, last_addr;
pgprot_t prot;
@@ -61,9 +56,10 @@ void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned l
t_addr = __va(phys_addr);
t_end = t_addr + (size - 1);
-
- for(page = virt_to_page(t_addr); page <= virt_to_page(t_end); page++)
- if(!PageReserved(page))
+
+ for (page = virt_to_page(t_addr);
+ page <= virt_to_page(t_end); page++)
+ if (!PageReserved(page))
return NULL;
}
@@ -85,7 +81,7 @@ void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned l
area->phys_addr = phys_addr;
addr = (void __iomem *) area->addr;
if (ioremap_page_range((unsigned long) addr,
- (unsigned long) addr + size, phys_addr, prot)) {
+ (unsigned long) addr + size, phys_addr, prot)) {
vunmap((void __force *) addr);
return NULL;
}
@@ -102,31 +98,31 @@ EXPORT_SYMBOL(__ioremap);
* make bus memory CPU accessible via the readb/readw/readl/writeb/
* writew/writel functions and the other mmio helpers. The returned
* address is not guaranteed to be usable directly as a virtual
- * address.
+ * address.
*
* This version of ioremap ensures that the memory is marked uncachable
* on the CPU as well as honouring existing caching rules from things like
- * the PCI bus. Note that there are other caches and buffers on many
+ * the PCI bus. Note that there are other caches and buffers on many
* busses. In particular driver authors should read up on PCI writes
*
* It's useful if some control registers are in such an area and
* write combining or read caching is not desirable:
- *
+ *
* Must be freed with iounmap.
*/
-
-void __iomem *ioremap_nocache (unsigned long phys_addr, unsigned long size)
+void __iomem *ioremap_nocache(unsigned long phys_addr, unsigned long size)
{
unsigned long last_addr;
void __iomem *p = __ioremap(phys_addr, size, _PAGE_PCD | _PAGE_PWT);
- if (!p)
- return p;
+
+ if (!p)
+ return p;
/* Guaranteed to be > phys_addr, as per __ioremap() */
last_addr = phys_addr + size - 1;
if (last_addr < virt_to_phys(high_memory) - 1) {
- struct page *ppage = virt_to_page(__va(phys_addr));
+ struct page *ppage = virt_to_page(__va(phys_addr));
unsigned long npages;
phys_addr &= PAGE_MASK;
@@ -135,18 +131,18 @@ void __iomem *ioremap_nocache (unsigned long phys_addr, unsigned long size)
last_addr = PAGE_ALIGN(last_addr);
/* .. but that's ok, because modulo-2**n arithmetic will make
- * the page-aligned "last - first" come out right.
- */
+ * the page-aligned "last - first" come out right.
+ */
npages = (last_addr - phys_addr) >> PAGE_SHIFT;
- if (change_page_attr(ppage, npages, PAGE_KERNEL_NOCACHE) < 0) {
- iounmap(p);
+ if (change_page_attr(ppage, npages, PAGE_KERNEL_NOCACHE) < 0) {
+ iounmap(p);
p = NULL;
}
global_flush_tlb();
}
- return p;
+ return p;
}
EXPORT_SYMBOL(ioremap_nocache);
@@ -169,10 +165,11 @@ void iounmap(volatile void __iomem *addr)
* of ISA space. So handle that here.
*/
if (addr >= phys_to_virt(ISA_START_ADDRESS) &&
- addr < phys_to_virt(ISA_END_ADDRESS))
+ addr < phys_to_virt(ISA_END_ADDRESS))
return;
- addr = (volatile void __iomem *)(PAGE_MASK & (unsigned long __force)addr);
+ addr = (volatile void __iomem *)
+ (PAGE_MASK & (unsigned long __force)addr);
/* Use the vm area unlocked, assuming the caller
ensures there isn't another iounmap for the same address
@@ -187,7 +184,7 @@ void iounmap(volatile void __iomem *addr)
read_unlock(&vmlist_lock);
if (!p) {
- printk("iounmap: bad address %p\n", addr);
+ printk(KERN_ERR "iounmap: bad address %p\n", addr);
dump_stack();
return;
}
@@ -198,12 +195,12 @@ void iounmap(volatile void __iomem *addr)
get_vm_area_size(p) >> PAGE_SHIFT,
PAGE_KERNEL);
global_flush_tlb();
- }
+ }
/* Finally remove it */
o = remove_vm_area((void *)addr);
BUG_ON(p != o || o == NULL);
- kfree(p);
+ kfree(p);
}
EXPORT_SYMBOL(iounmap);
@@ -237,7 +234,7 @@ void __init early_ioremap_init(void)
unsigned long *pgd;
if (early_ioremap_debug)
- printk("early_ioremap_init()\n");
+ printk(KERN_DEBUG "early_ioremap_init()\n");
pgd = early_ioremap_pgd(fix_to_virt(FIX_BTMAP_BEGIN));
*pgd = __pa(bm_pte) | _PAGE_TABLE;
@@ -248,15 +245,16 @@ void __init early_ioremap_init(void)
*/
if (pgd != early_ioremap_pgd(fix_to_virt(FIX_BTMAP_END))) {
WARN_ON(1);
- printk("pgd %p != %p\n",
- pgd, early_ioremap_pgd(fix_to_virt(FIX_BTMAP_END)));
- printk("fix_to_virt(FIX_BTMAP_BEGIN): %08lx\n",
- fix_to_virt(FIX_BTMAP_BEGIN));
- printk("fix_to_virt(FIX_BTMAP_END): %08lx\n",
- fix_to_virt(FIX_BTMAP_END));
-
- printk("FIX_BTMAP_END: %d\n", FIX_BTMAP_END);
- printk("FIX_BTMAP_BEGIN: %d\n", FIX_BTMAP_BEGIN);
+ printk(KERN_WARNING "pgd %p != %p\n",
+ pgd, early_ioremap_pgd(fix_to_virt(FIX_BTMAP_END)));
+ printk(KERN_WARNING "fix_to_virt(FIX_BTMAP_BEGIN): %08lx\n",
+ fix_to_virt(FIX_BTMAP_BEGIN));
+ printk(KERN_WARNING "fix_to_virt(FIX_BTMAP_END): %08lx\n",
+ fix_to_virt(FIX_BTMAP_END));
+
+ printk(KERN_WARNING "FIX_BTMAP_END: %d\n", FIX_BTMAP_END);
+ printk(KERN_WARNING "FIX_BTMAP_BEGIN: %d\n",
+ FIX_BTMAP_BEGIN);
}
}
@@ -265,7 +263,7 @@ void __init early_ioremap_clear(void)
unsigned long *pgd;
if (early_ioremap_debug)
- printk("early_ioremap_clear()\n");
+ printk(KERN_DEBUG "early_ioremap_clear()\n");
pgd = early_ioremap_pgd(fix_to_virt(FIX_BTMAP_BEGIN));
*pgd = 0;
@@ -331,10 +329,10 @@ static int __init check_early_ioremap_leak(void)
return 0;
printk(KERN_WARNING
- "Debug warning: early ioremap leak of %d areas detected.\n",
- early_ioremap_nested);
+ "Debug warning: early ioremap leak of %d areas detected.\n",
+ early_ioremap_nested);
printk(KERN_WARNING
- "please boot with early_ioremap_debug and report the dmesg.\n");
+ "please boot with early_ioremap_debug and report the dmesg.\n");
WARN_ON(1);
return 1;
@@ -351,8 +349,8 @@ void __init *early_ioremap(unsigned long phys_addr, unsigned long size)
nesting = early_ioremap_nested;
if (early_ioremap_debug) {
- printk("early_ioremap(%08lx, %08lx) [%d] => ",
- phys_addr, size, nesting);
+ printk(KERN_DEBUG "early_ioremap(%08lx, %08lx) [%d] => ",
+ phys_addr, size, nesting);
dump_stack();
}
@@ -398,7 +396,7 @@ void __init *early_ioremap(unsigned long phys_addr, unsigned long size)
if (early_ioremap_debug)
printk(KERN_CONT "%08lx + %08lx\n", offset, fix_to_virt(idx0));
- return (void*) (offset + fix_to_virt(idx0));
+ return (void *) (offset + fix_to_virt(idx0));
}
void __init early_iounmap(void *addr, unsigned long size)
@@ -413,7 +411,8 @@ void __init early_iounmap(void *addr, unsigned long size)
WARN_ON(nesting < 0);
if (early_ioremap_debug) {
- printk("early_iounmap(%p, %08lx) [%d]\n", addr, size, nesting);
+ printk(KERN_DEBUG "early_iounmap(%p, %08lx) [%d]\n", addr,
+ size, nesting);
dump_stack();
}
diff --git a/arch/x86/mm/ioremap_64.c b/arch/x86/mm/ioremap_64.c
index a37556124c86..bd962cc636c5 100644
--- a/arch/x86/mm/ioremap_64.c
+++ b/arch/x86/mm/ioremap_64.c
@@ -1,6 +1,4 @@
/*
- * arch/x86_64/mm/ioremap.c
- *
* Re-map IO memory to kernel address space so that we can access it.
* This is needed for high PCI addresses that aren't mapped in the
* 640k-1MB IO memory area on PC's
@@ -33,9 +31,8 @@ EXPORT_SYMBOL(__phys_addr);
* Fix up the linear direct mapping of the kernel to avoid cache attribute
* conflicts.
*/
-static int
-ioremap_change_attr(unsigned long phys_addr, unsigned long size,
- unsigned long flags)
+static int ioremap_change_attr(unsigned long phys_addr, unsigned long size,
+ unsigned long flags)
{
int err = 0;
if (phys_addr + size - 1 < (end_pfn_map << PAGE_SHIFT)) {
@@ -50,10 +47,12 @@ ioremap_change_attr(unsigned long phys_addr, unsigned long size,
if (!lookup_address(vaddr, &level))
return err;
/*
- * Must use a address here and not struct page because the phys addr
- * can be a in hole between nodes and not have an memmap entry.
+ * Must use a address here and not struct page because
+ * the phys addr can be a in hole between nodes and
+ * not have an memmap entry.
*/
- err = change_page_attr_addr(vaddr,npages,MAKE_GLOBAL(__PAGE_KERNEL|flags));
+ err = change_page_attr_addr(vaddr,npages,
+ MAKE_GLOBAL(__PAGE_KERNEL|flags));
if (!err)
global_flush_tlb();
}
@@ -61,10 +60,6 @@ ioremap_change_attr(unsigned long phys_addr, unsigned long size,
}
/*
- * Generic mapping function
- */
-
-/*
* Remap an arbitrary physical address space into the kernel virtual
* address space. Needed when the kernel wants to access high addresses
* directly.
@@ -73,10 +68,11 @@ ioremap_change_attr(unsigned long phys_addr, unsigned long size,
* have to convert them into an offset in a page-aligned mapping, but the
* caller shouldn't need to know that small detail.
*/
-void __iomem * __ioremap(unsigned long phys_addr, unsigned long size, unsigned long flags)
+void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
+ unsigned long flags)
{
- void * addr;
- struct vm_struct * area;
+ void *addr;
+ struct vm_struct *area;
unsigned long offset, last_addr;
pgprot_t pgprot;
@@ -130,20 +126,19 @@ EXPORT_SYMBOL(__ioremap);
* make bus memory CPU accessible via the readb/readw/readl/writeb/
* writew/writel functions and the other mmio helpers. The returned
* address is not guaranteed to be usable directly as a virtual
- * address.
+ * address.
*
* This version of ioremap ensures that the memory is marked uncachable
* on the CPU as well as honouring existing caching rules from things like
- * the PCI bus. Note that there are other caches and buffers on many
+ * the PCI bus. Note that there are other caches and buffers on many
* busses. In particular driver authors should read up on PCI writes
*
* It's useful if some control registers are in such an area and
* write combining or read caching is not desirable:
- *
+ *
* Must be freed with iounmap.
*/
-
-void __iomem *ioremap_nocache (unsigned long phys_addr, unsigned long size)
+void __iomem *ioremap_nocache(unsigned long phys_addr, unsigned long size)
{
return __ioremap(phys_addr, size, _PAGE_PCD | _PAGE_PWT);
}
@@ -159,13 +154,14 @@ void iounmap(volatile void __iomem *addr)
{
struct vm_struct *p, *o;
- if (addr <= high_memory)
- return;
+ if (addr <= high_memory)
+ return;
if (addr >= phys_to_virt(ISA_START_ADDRESS) &&
- addr < phys_to_virt(ISA_END_ADDRESS))
+ addr < phys_to_virt(ISA_END_ADDRESS))
return;
- addr = (volatile void __iomem *)(PAGE_MASK & (unsigned long __force)addr);
+ addr = (volatile void __iomem *)
+ (PAGE_MASK & (unsigned long __force)addr);
/* Use the vm area unlocked, assuming the caller
ensures there isn't another iounmap for the same address
in parallel. Reuse of the virtual address is prevented by
@@ -179,7 +175,7 @@ void iounmap(volatile void __iomem *addr)
read_unlock(&vmlist_lock);
if (!p) {
- printk("iounmap: bad address %p\n", addr);
+ printk(KERN_ERR "iounmap: bad address %p\n", addr);
dump_stack();
return;
}
@@ -191,7 +187,7 @@ void iounmap(volatile void __iomem *addr)
/* Finally remove it */
o = remove_vm_area((void *)addr);
BUG_ON(p != o || o == NULL);
- kfree(p);
+ kfree(p);
}
EXPORT_SYMBOL(iounmap);