/* * * Copyright (c) 2009, Microsoft Corporation. * * This program is free software; you can redistribute it and/or modify it * under the terms and conditions of the GNU General Public License, * version 2, as published by the Free Software Foundation. * * This program is distributed in the hope it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for * more details. * * You should have received a copy of the GNU General Public License along with * this program; if not, write to the Free Software Foundation, Inc., 59 Temple * Place - Suite 330, Boston, MA 02111-1307 USA. * * Authors: * Haiyang Zhang * Hank Janssen * */ #define KERNEL_2_6_27 #include #include #include #include #include #include //#include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "include/osd.h" // // Data types // typedef struct _TIMER { struct timer_list timer; PFN_TIMER_CALLBACK callback; void* context; }TIMER; typedef struct _WAITEVENT { int condition; wait_queue_head_t event; } WAITEVENT; typedef struct _SPINLOCK { spinlock_t lock; unsigned long flags; } SPINLOCK; typedef struct _WORKQUEUE { struct workqueue_struct *queue; } WORKQUEUE; typedef struct _WORKITEM { struct work_struct work; PFN_WORKITEM_CALLBACK callback; void* context; } WORKITEM; // // Global // void LogMsg(const char *fmt, ...) { #ifdef KERNEL_2_6_5 char buf[1024]; #endif va_list args; va_start(args, fmt); #ifdef KERNEL_2_6_5 vsnprintf(buf, 1024, fmt, args); va_end(args); printk(buf); #else vprintk(fmt, args); va_end(args); #endif } void BitSet(unsigned int* addr, int bit) { set_bit(bit, (unsigned long*)addr); } int BitTest(unsigned int* addr, int bit) { return test_bit(bit, (unsigned long*)addr); } void BitClear(unsigned int* addr, int bit) { clear_bit(bit, (unsigned long*)addr); } int BitTestAndClear(unsigned int* addr, int bit) { return test_and_clear_bit(bit, (unsigned long*)addr); } int BitTestAndSet(unsigned int* addr, int bit) { return test_and_set_bit(bit, (unsigned long*)addr); } int InterlockedIncrement(int *val) { #ifdef KERNEL_2_6_5 int i; local_irq_disable(); i = atomic_read((atomic_t*)val); atomic_set((atomic_t*)val, i+1); local_irq_enable(); return i+1; #else return atomic_inc_return((atomic_t*)val); #endif } int InterlockedDecrement(int *val) { #ifdef KERNEL_2_6_5 int i; local_irq_disable(); i = atomic_read((atomic_t*)val); atomic_set((atomic_t*)val, i-1); local_irq_enable(); return i-1; #else return atomic_dec_return((atomic_t*)val); #endif } #ifndef atomic_cmpxchg #define atomic_cmpxchg(v, old, new) ((int)cmpxchg(&((v)->counter), old, new)) #endif int InterlockedCompareExchange(int *val, int new, int curr) { //return ((int)cmpxchg(((atomic_t*)val), curr, new)); return atomic_cmpxchg((atomic_t*)val, curr, new); } void Sleep(unsigned long usecs) { udelay(usecs); } void* VirtualAllocExec(unsigned int size) { #ifdef __x86_64__ return __vmalloc(size, GFP_KERNEL, PAGE_KERNEL_EXEC); #else return __vmalloc(size, GFP_KERNEL, __pgprot(__PAGE_KERNEL & (~_PAGE_NX))); #endif } void VirtualFree(void* VirtAddr) { return vfree(VirtAddr); } void* PageAlloc(unsigned int count) { void *p; p = (void *)__get_free_pages(GFP_KERNEL, get_order(count * PAGE_SIZE)); if (p) memset(p, 0, count * PAGE_SIZE); return p; //struct page* page = alloc_page(GFP_KERNEL|__GFP_ZERO); //void *p; ////BUGBUG: We need to use kmap in case we are in HIMEM region //p = page_address(page); //if (p) memset(p, 0, PAGE_SIZE); //return p; } void PageFree(void* page, unsigned int count) { free_pages((unsigned long)page, get_order(count * PAGE_SIZE)); /*struct page* p = virt_to_page(page); __free_page(p);*/ } void* PageMapVirtualAddress(unsigned long Pfn) { return kmap_atomic(pfn_to_page(Pfn), KM_IRQ0); } void PageUnmapVirtualAddress(void* VirtAddr) { kunmap_atomic(VirtAddr, KM_IRQ0); } void* MemAlloc(unsigned int size) { return kmalloc(size, GFP_KERNEL); } void* MemAllocZeroed(unsigned int size) { void *p = kmalloc(size, GFP_KERNEL); if (p) memset(p, 0, size); return p; } void* MemAllocAtomic(unsigned int size) { return kmalloc(size, GFP_ATOMIC); } void MemFree(void* buf) { kfree(buf); } void *MemMapIO(unsigned long phys, unsigned long size) { #if X2V_LINUX #ifdef __x86_64__ return (void*)(phys + 0xFFFF83000C000000); #else // i386 return (void*)(phys + 0xfb000000); #endif #else return (void*)GetVirtualAddress(phys); //return ioremap_nocache(phys, size); #endif } void MemUnmapIO(void *virt) { //iounmap(virt); } void MemoryFence() { mb(); } void TimerCallback(unsigned long data) { TIMER* t = (TIMER*)data; t->callback(t->context); } HANDLE TimerCreate(PFN_TIMER_CALLBACK pfnTimerCB, void* context) { TIMER* t = kmalloc(sizeof(TIMER), GFP_KERNEL); if (!t) { return NULL; } t->callback = pfnTimerCB; t->context = context; init_timer(&t->timer); t->timer.data = (unsigned long)t; t->timer.function = TimerCallback; return t; } void TimerStart(HANDLE hTimer, u32 expirationInUs) { TIMER* t = (TIMER* )hTimer; t->timer.expires = jiffies + usecs_to_jiffies(expirationInUs); add_timer(&t->timer); } int TimerStop(HANDLE hTimer) { TIMER* t = (TIMER* )hTimer; return del_timer(&t->timer); } void TimerClose(HANDLE hTimer) { TIMER* t = (TIMER* )hTimer; del_timer(&t->timer); kfree(t); } SIZE_T GetTickCount(void) { return jiffies; } signed long long GetTimestamp(void) { struct timeval t; do_gettimeofday(&t); return timeval_to_ns(&t); } HANDLE WaitEventCreate(void) { WAITEVENT* wait = kmalloc(sizeof(WAITEVENT), GFP_KERNEL); if (!wait) { return NULL; } wait->condition = 0; init_waitqueue_head(&wait->event); return wait; } void WaitEventClose(HANDLE hWait) { WAITEVENT* waitEvent = (WAITEVENT* )hWait; kfree(waitEvent); } void WaitEventSet(HANDLE hWait) { WAITEVENT* waitEvent = (WAITEVENT* )hWait; waitEvent->condition = 1; wake_up_interruptible(&waitEvent->event); } int WaitEventWait(HANDLE hWait) { int ret=0; WAITEVENT* waitEvent = (WAITEVENT* )hWait; ret= wait_event_interruptible(waitEvent->event, waitEvent->condition); waitEvent->condition = 0; return ret; } int WaitEventWaitEx(HANDLE hWait, u32 TimeoutInMs) { int ret=0; WAITEVENT* waitEvent = (WAITEVENT* )hWait; ret= wait_event_interruptible_timeout(waitEvent->event, waitEvent->condition, msecs_to_jiffies(TimeoutInMs)); waitEvent->condition = 0; return ret; } HANDLE SpinlockCreate(void) { SPINLOCK* spin = kmalloc(sizeof(SPINLOCK), GFP_KERNEL); if (!spin) { return NULL; } spin_lock_init(&spin->lock); return spin; } void SpinlockAcquire(HANDLE hSpin) { SPINLOCK* spin = (SPINLOCK* )hSpin; spin_lock_irqsave(&spin->lock, spin->flags); } void SpinlockRelease(HANDLE hSpin) { SPINLOCK* spin = (SPINLOCK* )hSpin; spin_unlock_irqrestore(&spin->lock, spin->flags); } void SpinlockClose(HANDLE hSpin) { SPINLOCK* spin = (SPINLOCK* )hSpin; kfree(spin); } void* Physical2LogicalAddr(ULONG_PTR PhysAddr) { void* logicalAddr = phys_to_virt(PhysAddr); BUG_ON(!virt_addr_valid(logicalAddr)); return logicalAddr; } ULONG_PTR Logical2PhysicalAddr(void * LogicalAddr) { BUG_ON(!virt_addr_valid(LogicalAddr)); return virt_to_phys(LogicalAddr); } ULONG_PTR Virtual2Physical(void * VirtAddr) { ULONG_PTR pfn = vmalloc_to_pfn(VirtAddr); return pfn << PAGE_SHIFT; } #ifdef KERNEL_2_6_27 void WorkItemCallback(struct work_struct *work) #else void WorkItemCallback(void* work) #endif { WORKITEM* w = (WORKITEM*)work; w->callback(w->context); kfree(w); } HANDLE WorkQueueCreate(char* name) { WORKQUEUE *wq = kmalloc(sizeof(WORKQUEUE), GFP_KERNEL); if (!wq) { return NULL; } wq->queue = create_workqueue(name); return wq; } void WorkQueueClose(HANDLE hWorkQueue) { WORKQUEUE *wq = (WORKQUEUE *)hWorkQueue; destroy_workqueue(wq->queue); return; } int WorkQueueQueueWorkItem(HANDLE hWorkQueue, PFN_WORKITEM_CALLBACK workItem, void* context) { WORKQUEUE *wq = (WORKQUEUE *)hWorkQueue; WORKITEM* w = kmalloc(sizeof(WORKITEM), GFP_ATOMIC); if (!w) { return -1; } w->callback = workItem, w->context = context; #ifdef KERNEL_2_6_27 INIT_WORK(&w->work, WorkItemCallback); #else INIT_WORK(&w->work, WorkItemCallback, w); #endif return queue_work(wq->queue, &w->work); } void QueueWorkItem(PFN_WORKITEM_CALLBACK workItem, void* context) { WORKITEM* w = kmalloc(sizeof(WORKITEM), GFP_ATOMIC); if (!w) { return; } w->callback = workItem, w->context = context; #ifdef KERNEL_2_6_27 INIT_WORK(&w->work, WorkItemCallback); #else INIT_WORK(&w->work, WorkItemCallback, w); #endif schedule_work(&w->work); }