*/
void fastcall __sched mutex_lock(struct mutex *lock)
{
+ might_sleep();
/*
* The locking fastpath is the 1->0 transition from
* 'unlocked' into 'locked' state.
static fastcall noinline void
__mutex_unlock_slowpath(atomic_t *lock_count __IP_DECL__)
{
- struct mutex *lock = container_of(lock_count, struct mutex, count);
+ struct mutex *lock = container_of(lock_count, struct mutex, count);
DEBUG_WARN_ON(lock->owner != current_thread_info());
*/
int fastcall __sched mutex_lock_interruptible(struct mutex *lock)
{
+ might_sleep();
return __mutex_fastpath_lock_retval
(&lock->count, __mutex_lock_interruptible_slowpath);
}
}
EXPORT_SYMBOL(mutex_trylock);
-
-
-