/* * Copyright (C) 2004-2006 Atmel Corporation * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */ #include #include #include #include #include #include #include static inline int remap_area_pte(pte_t *pte, unsigned long address, unsigned long end, unsigned long phys_addr, pgprot_t prot) { unsigned long pfn; pfn = phys_addr >> PAGE_SHIFT; do { WARN_ON(!pte_none(*pte)); set_pte(pte, pfn_pte(pfn, prot)); address += PAGE_SIZE; pfn++; pte++; } while (address && (address < end)); return 0; } static inline int remap_area_pmd(pmd_t *pmd, unsigned long address, unsigned long end, unsigned long phys_addr, pgprot_t prot) { unsigned long next; phys_addr -= address; do { pte_t *pte = pte_alloc_kernel(pmd, address); if (!pte) return -ENOMEM; next = (address + PMD_SIZE) & PMD_MASK; if (remap_area_pte(pte, address, next, address + phys_addr, prot)) return -ENOMEM; address = next; pmd++; } while (address && (address < end)); return 0; } static int remap_area_pud(pud_t *pud, unsigned long address, unsigned long end, unsigned long phys_addr, pgprot_t prot) { unsigned long next; phys_addr -= address; do { pmd_t *pmd = pmd_alloc(&init_mm, pud, address); if (!pmd) return -ENOMEM; next = (address + PUD_SIZE) & PUD_MASK; if (remap_area_pmd(pmd, address, next, phys_addr + address, prot)) return -ENOMEM; address = next; pud++; } while (address && address < end); return 0; } static int remap_area_pages(unsigned long address, unsigned long phys_addr, size_t size, pgprot_t prot) { unsigned long end = address + size; unsigned long next; pgd_t *pgd; int err = 0; phys_addr -= address; pgd = pgd_offset_k(address); flush_cache_all(); BUG_ON(address >= end); spin_lock(&init_mm.page_table_lock); do { pud_t *pud = pud_alloc(&init_mm, pgd, address); err = -ENOMEM; if (!pud) break; next = (address + PGDIR_SIZE) & PGDIR_MASK; if (next < address || next > end) next = end; err = remap_area_pud(pud, address, next, phys_addr + address, prot); if (err) break; address = next; pgd++; } while (address && (address < end)); spin_unlock(&init_mm.page_table_lock); flush_tlb_all(); return err; } /* * Re-map an arbitrary physical address space into the kernel virtual * address space. Needed when the kernel wants to access physical * memory directly. */ void __iomem *__ioremap(unsigned long phys_addr, size_t size, unsigned long flags) { void *addr; struct vm_struct *area; unsigned long offset, last_addr; pgprot_t prot; /* * Check if we can simply use the P4 segment. This area is * uncacheable, so if caching/buffering is requested, we can't * use it. */ if ((phys_addr >= P4SEG) && (flags == 0)) return (void __iomem *)phys_addr; /* Don't allow wraparound or zero size */ last_addr = phys_addr + size - 1; if (!size || last_addr < phys_addr) return NULL; /* * XXX: When mapping regular RAM, we'd better make damn sure * it's never used for anything else. But this is really the * caller's responsibility... */ if (PHYSADDR(P2SEGADDR(phys_addr)) == phys_addr) return (void __iomem *)P2SEGADDR(phys_addr); /* Mappings have to be page-aligned */ offset = phys_addr & ~PAGE_MASK; phys_addr &= PAGE_MASK; size = PAGE_ALIGN(last_addr + 1) - phys_addr; prot = __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_TYPE_SMALL | flags); /* * Ok, go for it.. */ area = get_vm_area(size, VM_IOREMAP); if (!area) return NULL; area->phys_addr = phys_addr; addr = area->addr; if (remap_area_pages((unsigned long)addr, phys_addr, size, prot)) { vunmap(addr); return NULL; } return (void __iomem *)(offset + (char *)addr); } EXPORT_SYMBOL(__ioremap); void __iounmap(void __iomem *addr) { struct vm_struct *p; if ((unsigned long)addr >= P4SEG) return; p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr)); if (unlikely(!p)) { printk (KERN_ERR "iounmap: bad address %p\n", addr); return; } kfree (p); } EXPORT_SYMBOL(__iounmap);