diff options
Diffstat (limited to 'kernel/rtmutex.c')
| -rw-r--r-- | kernel/rtmutex.c | 106 | 
1 files changed, 53 insertions, 53 deletions
| diff --git a/kernel/rtmutex.c b/kernel/rtmutex.c index 29bd4baf9e75..a9604815786a 100644 --- a/kernel/rtmutex.c +++ b/kernel/rtmutex.c @@ -138,9 +138,9 @@ static void rt_mutex_adjust_prio(struct task_struct *task)  {  	unsigned long flags; -	spin_lock_irqsave(&task->pi_lock, flags); +	raw_spin_lock_irqsave(&task->pi_lock, flags);  	__rt_mutex_adjust_prio(task); -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  }  /* @@ -195,7 +195,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	/*  	 * Task can not go away as we did a get_task() before !  	 */ -	spin_lock_irqsave(&task->pi_lock, flags); +	raw_spin_lock_irqsave(&task->pi_lock, flags);  	waiter = task->pi_blocked_on;  	/* @@ -231,8 +231,8 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  		goto out_unlock_pi;  	lock = waiter->lock; -	if (!spin_trylock(&lock->wait_lock)) { -		spin_unlock_irqrestore(&task->pi_lock, flags); +	if (!raw_spin_trylock(&lock->wait_lock)) { +		raw_spin_unlock_irqrestore(&task->pi_lock, flags);  		cpu_relax();  		goto retry;  	} @@ -240,7 +240,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	/* Deadlock detection */  	if (lock == orig_lock || rt_mutex_owner(lock) == top_task) {  		debug_rt_mutex_deadlock(deadlock_detect, orig_waiter, lock); -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		ret = deadlock_detect ? -EDEADLK : 0;  		goto out_unlock_pi;  	} @@ -253,13 +253,13 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	plist_add(&waiter->list_entry, &lock->wait_list);  	/* Release the task */ -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	put_task_struct(task);  	/* Grab the next task */  	task = rt_mutex_owner(lock);  	get_task_struct(task); -	spin_lock_irqsave(&task->pi_lock, flags); +	raw_spin_lock_irqsave(&task->pi_lock, flags);  	if (waiter == rt_mutex_top_waiter(lock)) {  		/* Boost the owner */ @@ -277,10 +277,10 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  		__rt_mutex_adjust_prio(task);  	} -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	top_waiter = rt_mutex_top_waiter(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	if (!detect_deadlock && waiter != top_waiter)  		goto out_put_task; @@ -288,7 +288,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	goto again;   out_unlock_pi: -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);   out_put_task:  	put_task_struct(task); @@ -313,9 +313,9 @@ static inline int try_to_steal_lock(struct rt_mutex *lock,  	if (pendowner == task)  		return 1; -	spin_lock_irqsave(&pendowner->pi_lock, flags); +	raw_spin_lock_irqsave(&pendowner->pi_lock, flags);  	if (task->prio >= pendowner->prio) { -		spin_unlock_irqrestore(&pendowner->pi_lock, flags); +		raw_spin_unlock_irqrestore(&pendowner->pi_lock, flags);  		return 0;  	} @@ -325,7 +325,7 @@ static inline int try_to_steal_lock(struct rt_mutex *lock,  	 * priority.  	 */  	if (likely(!rt_mutex_has_waiters(lock))) { -		spin_unlock_irqrestore(&pendowner->pi_lock, flags); +		raw_spin_unlock_irqrestore(&pendowner->pi_lock, flags);  		return 1;  	} @@ -333,7 +333,7 @@ static inline int try_to_steal_lock(struct rt_mutex *lock,  	next = rt_mutex_top_waiter(lock);  	plist_del(&next->pi_list_entry, &pendowner->pi_waiters);  	__rt_mutex_adjust_prio(pendowner); -	spin_unlock_irqrestore(&pendowner->pi_lock, flags); +	raw_spin_unlock_irqrestore(&pendowner->pi_lock, flags);  	/*  	 * We are going to steal the lock and a waiter was @@ -350,10 +350,10 @@ static inline int try_to_steal_lock(struct rt_mutex *lock,  	 * might be task:  	 */  	if (likely(next->task != task)) { -		spin_lock_irqsave(&task->pi_lock, flags); +		raw_spin_lock_irqsave(&task->pi_lock, flags);  		plist_add(&next->pi_list_entry, &task->pi_waiters);  		__rt_mutex_adjust_prio(task); -		spin_unlock_irqrestore(&task->pi_lock, flags); +		raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	}  	return 1;  } @@ -420,7 +420,7 @@ static int task_blocks_on_rt_mutex(struct rt_mutex *lock,  	unsigned long flags;  	int chain_walk = 0, res; -	spin_lock_irqsave(&task->pi_lock, flags); +	raw_spin_lock_irqsave(&task->pi_lock, flags);  	__rt_mutex_adjust_prio(task);  	waiter->task = task;  	waiter->lock = lock; @@ -434,17 +434,17 @@ static int task_blocks_on_rt_mutex(struct rt_mutex *lock,  	task->pi_blocked_on = waiter; -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	if (waiter == rt_mutex_top_waiter(lock)) { -		spin_lock_irqsave(&owner->pi_lock, flags); +		raw_spin_lock_irqsave(&owner->pi_lock, flags);  		plist_del(&top_waiter->pi_list_entry, &owner->pi_waiters);  		plist_add(&waiter->pi_list_entry, &owner->pi_waiters);  		__rt_mutex_adjust_prio(owner);  		if (owner->pi_blocked_on)  			chain_walk = 1; -		spin_unlock_irqrestore(&owner->pi_lock, flags); +		raw_spin_unlock_irqrestore(&owner->pi_lock, flags);  	}  	else if (debug_rt_mutex_detect_deadlock(waiter, detect_deadlock))  		chain_walk = 1; @@ -459,12 +459,12 @@ static int task_blocks_on_rt_mutex(struct rt_mutex *lock,  	 */  	get_task_struct(owner); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	res = rt_mutex_adjust_prio_chain(owner, detect_deadlock, lock, waiter,  					 task); -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	return res;  } @@ -483,7 +483,7 @@ static void wakeup_next_waiter(struct rt_mutex *lock)  	struct task_struct *pendowner;  	unsigned long flags; -	spin_lock_irqsave(¤t->pi_lock, flags); +	raw_spin_lock_irqsave(¤t->pi_lock, flags);  	waiter = rt_mutex_top_waiter(lock);  	plist_del(&waiter->list_entry, &lock->wait_list); @@ -500,7 +500,7 @@ static void wakeup_next_waiter(struct rt_mutex *lock)  	rt_mutex_set_owner(lock, pendowner, RT_MUTEX_OWNER_PENDING); -	spin_unlock_irqrestore(¤t->pi_lock, flags); +	raw_spin_unlock_irqrestore(¤t->pi_lock, flags);  	/*  	 * Clear the pi_blocked_on variable and enqueue a possible @@ -509,7 +509,7 @@ static void wakeup_next_waiter(struct rt_mutex *lock)  	 * waiter with higher priority than pending-owner->normal_prio  	 * is blocked on the unboosted (pending) owner.  	 */ -	spin_lock_irqsave(&pendowner->pi_lock, flags); +	raw_spin_lock_irqsave(&pendowner->pi_lock, flags);  	WARN_ON(!pendowner->pi_blocked_on);  	WARN_ON(pendowner->pi_blocked_on != waiter); @@ -523,7 +523,7 @@ static void wakeup_next_waiter(struct rt_mutex *lock)  		next = rt_mutex_top_waiter(lock);  		plist_add(&next->pi_list_entry, &pendowner->pi_waiters);  	} -	spin_unlock_irqrestore(&pendowner->pi_lock, flags); +	raw_spin_unlock_irqrestore(&pendowner->pi_lock, flags);  	wake_up_process(pendowner);  } @@ -541,15 +541,15 @@ static void remove_waiter(struct rt_mutex *lock,  	unsigned long flags;  	int chain_walk = 0; -	spin_lock_irqsave(¤t->pi_lock, flags); +	raw_spin_lock_irqsave(¤t->pi_lock, flags);  	plist_del(&waiter->list_entry, &lock->wait_list);  	waiter->task = NULL;  	current->pi_blocked_on = NULL; -	spin_unlock_irqrestore(¤t->pi_lock, flags); +	raw_spin_unlock_irqrestore(¤t->pi_lock, flags);  	if (first && owner != current) { -		spin_lock_irqsave(&owner->pi_lock, flags); +		raw_spin_lock_irqsave(&owner->pi_lock, flags);  		plist_del(&waiter->pi_list_entry, &owner->pi_waiters); @@ -564,7 +564,7 @@ static void remove_waiter(struct rt_mutex *lock,  		if (owner->pi_blocked_on)  			chain_walk = 1; -		spin_unlock_irqrestore(&owner->pi_lock, flags); +		raw_spin_unlock_irqrestore(&owner->pi_lock, flags);  	}  	WARN_ON(!plist_node_empty(&waiter->pi_list_entry)); @@ -575,11 +575,11 @@ static void remove_waiter(struct rt_mutex *lock,  	/* gets dropped in rt_mutex_adjust_prio_chain()! */  	get_task_struct(owner); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	rt_mutex_adjust_prio_chain(owner, 0, lock, NULL, current); -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  }  /* @@ -592,15 +592,15 @@ void rt_mutex_adjust_pi(struct task_struct *task)  	struct rt_mutex_waiter *waiter;  	unsigned long flags; -	spin_lock_irqsave(&task->pi_lock, flags); +	raw_spin_lock_irqsave(&task->pi_lock, flags);  	waiter = task->pi_blocked_on;  	if (!waiter || waiter->list_entry.prio == task->prio) { -		spin_unlock_irqrestore(&task->pi_lock, flags); +		raw_spin_unlock_irqrestore(&task->pi_lock, flags);  		return;  	} -	spin_unlock_irqrestore(&task->pi_lock, flags); +	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	/* gets dropped in rt_mutex_adjust_prio_chain()! */  	get_task_struct(task); @@ -672,14 +672,14 @@ __rt_mutex_slowlock(struct rt_mutex *lock, int state,  				break;  		} -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		debug_rt_mutex_print_deadlock(waiter);  		if (waiter->task)  			schedule_rt_mutex(lock); -		spin_lock(&lock->wait_lock); +		raw_spin_lock(&lock->wait_lock);  		set_current_state(state);  	} @@ -700,11 +700,11 @@ rt_mutex_slowlock(struct rt_mutex *lock, int state,  	debug_rt_mutex_init_waiter(&waiter);  	waiter.task = NULL; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	/* Try to acquire the lock again: */  	if (try_to_take_rt_mutex(lock)) { -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		return 0;  	} @@ -731,7 +731,7 @@ rt_mutex_slowlock(struct rt_mutex *lock, int state,  	 */  	fixup_rt_mutex_waiters(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/* Remove pending timer: */  	if (unlikely(timeout)) @@ -758,7 +758,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  {  	int ret = 0; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	if (likely(rt_mutex_owner(lock) != current)) { @@ -770,7 +770,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  		fixup_rt_mutex_waiters(lock);  	} -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	return ret;  } @@ -781,7 +781,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  static void __sched  rt_mutex_slowunlock(struct rt_mutex *lock)  { -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	debug_rt_mutex_unlock(lock); @@ -789,13 +789,13 @@ rt_mutex_slowunlock(struct rt_mutex *lock)  	if (!rt_mutex_has_waiters(lock)) {  		lock->owner = NULL; -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		return;  	}  	wakeup_next_waiter(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/* Undo pi boosting if necessary: */  	rt_mutex_adjust_prio(current); @@ -970,8 +970,8 @@ EXPORT_SYMBOL_GPL(rt_mutex_destroy);  void __rt_mutex_init(struct rt_mutex *lock, const char *name)  {  	lock->owner = NULL; -	spin_lock_init(&lock->wait_lock); -	plist_head_init(&lock->wait_list, &lock->wait_lock); +	raw_spin_lock_init(&lock->wait_lock); +	plist_head_init_raw(&lock->wait_list, &lock->wait_lock);  	debug_rt_mutex_init(lock, name);  } @@ -1032,7 +1032,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  {  	int ret; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	mark_rt_mutex_waiters(lock); @@ -1040,7 +1040,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  		/* We got the lock for task. */  		debug_rt_mutex_lock(lock);  		rt_mutex_set_owner(lock, task, 0); -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		rt_mutex_deadlock_account_lock(lock, task);  		return 1;  	} @@ -1056,7 +1056,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  		 */  		ret = 0;  	} -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	debug_rt_mutex_print_deadlock(waiter); @@ -1106,7 +1106,7 @@ int rt_mutex_finish_proxy_lock(struct rt_mutex *lock,  {  	int ret; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	set_current_state(TASK_INTERRUPTIBLE); @@ -1124,7 +1124,7 @@ int rt_mutex_finish_proxy_lock(struct rt_mutex *lock,  	 */  	fixup_rt_mutex_waiters(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/*  	 * Readjust priority, when we did not get the lock. We might have been | 
