aboutsummaryrefslogtreecommitdiff
path: root/kernel
diff options
context:
space:
mode:
authorArjan van de Ven <arjan@linux.intel.com>2009-01-09 13:23:45 -0800
committerArjan van de Ven <arjan@linux.intel.com>2009-01-09 13:23:45 -0800
commitcdb80f630be5cbc23d82331f24dc4704f75b64f4 (patch)
tree537545e474b5906e4887aaa28e4e832134a97fa3 /kernel
parentb32714ba29358a688ef337d5297bf4bdc9f596dc (diff)
async: make async a command line option for now
... and have it default off. This does allow people to work with it for testing. Signed-off-by: Arjan van de Ven <arjan@linux.intel.com>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/async.c16
1 files changed, 14 insertions, 2 deletions
diff --git a/kernel/async.c b/kernel/async.c
index 64cc916299a..f286e9f2b73 100644
--- a/kernel/async.c
+++ b/kernel/async.c
@@ -65,6 +65,8 @@ static LIST_HEAD(async_pending);
static LIST_HEAD(async_running);
static DEFINE_SPINLOCK(async_lock);
+static int async_enabled = 0;
+
struct async_entry {
struct list_head list;
async_cookie_t cookie;
@@ -169,7 +171,7 @@ static async_cookie_t __async_schedule(async_func_ptr *ptr, void *data, struct l
* If we're out of memory or if there's too much work
* pending already, we execute synchronously.
*/
- if (!entry || atomic_read(&entry_count) > MAX_WORK) {
+ if (!async_enabled || !entry || atomic_read(&entry_count) > MAX_WORK) {
kfree(entry);
spin_lock_irqsave(&async_lock, flags);
newcookie = next_cookie++;
@@ -316,8 +318,18 @@ static int async_manager_thread(void *unused)
static int __init async_init(void)
{
- kthread_run(async_manager_thread, NULL, "async/mgr");
+ if (async_enabled)
+ kthread_run(async_manager_thread, NULL, "async/mgr");
return 0;
}
+static int __init setup_async(char *str)
+{
+ async_enabled = 1;
+ return 1;
+}
+
+__setup("fastboot", setup_async);
+
+
core_initcall(async_init);