Merge branch 'linus' into core/locking
Conflicts: fs/btrfs/locking.c
This commit is contained in:
@@ -1323,8 +1323,8 @@ static inline void update_load_sub(struct load_weight *lw, unsigned long dec)
|
||||
* slice expiry etc.
|
||||
*/
|
||||
|
||||
#define WEIGHT_IDLEPRIO 2
|
||||
#define WMULT_IDLEPRIO (1 << 31)
|
||||
#define WEIGHT_IDLEPRIO 3
|
||||
#define WMULT_IDLEPRIO 1431655765
|
||||
|
||||
/*
|
||||
* Nice levels are multiplicative, with a gentle 10% change for every
|
||||
@@ -2266,6 +2266,16 @@ static int try_to_wake_up(struct task_struct *p, unsigned int state, int sync)
|
||||
if (!sched_feat(SYNC_WAKEUPS))
|
||||
sync = 0;
|
||||
|
||||
if (!sync) {
|
||||
if (current->se.avg_overlap < sysctl_sched_migration_cost &&
|
||||
p->se.avg_overlap < sysctl_sched_migration_cost)
|
||||
sync = 1;
|
||||
} else {
|
||||
if (current->se.avg_overlap >= sysctl_sched_migration_cost ||
|
||||
p->se.avg_overlap >= sysctl_sched_migration_cost)
|
||||
sync = 0;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
if (sched_feat(LB_WAKEUP_UPDATE)) {
|
||||
struct sched_domain *sd;
|
||||
@@ -4440,7 +4450,7 @@ void __kprobes sub_preempt_count(int val)
|
||||
/*
|
||||
* Underflow?
|
||||
*/
|
||||
if (DEBUG_LOCKS_WARN_ON(val > preempt_count() - (!!kernel_locked())))
|
||||
if (DEBUG_LOCKS_WARN_ON(val > preempt_count()))
|
||||
return;
|
||||
/*
|
||||
* Is the spinlock portion underflowing?
|
||||
@@ -4752,8 +4762,8 @@ EXPORT_SYMBOL(default_wake_function);
|
||||
* started to run but is not in state TASK_RUNNING. try_to_wake_up() returns
|
||||
* zero in this (rare) case, and we handle it by continuing to scan the queue.
|
||||
*/
|
||||
static void __wake_up_common(wait_queue_head_t *q, unsigned int mode,
|
||||
int nr_exclusive, int sync, void *key)
|
||||
void __wake_up_common(wait_queue_head_t *q, unsigned int mode,
|
||||
int nr_exclusive, int sync, void *key)
|
||||
{
|
||||
wait_queue_t *curr, *next;
|
||||
|
||||
@@ -5191,7 +5201,7 @@ int can_nice(const struct task_struct *p, const int nice)
|
||||
* sys_setpriority is a more generic, but much slower function that
|
||||
* does similar things.
|
||||
*/
|
||||
asmlinkage long sys_nice(int increment)
|
||||
SYSCALL_DEFINE1(nice, int, increment)
|
||||
{
|
||||
long nice, retval;
|
||||
|
||||
@@ -5498,8 +5508,8 @@ do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
|
||||
* @policy: new policy.
|
||||
* @param: structure containing the new RT priority.
|
||||
*/
|
||||
asmlinkage long
|
||||
sys_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
|
||||
SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy,
|
||||
struct sched_param __user *, param)
|
||||
{
|
||||
/* negative values for policy are not valid */
|
||||
if (policy < 0)
|
||||
@@ -5513,7 +5523,7 @@ sys_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
|
||||
* @pid: the pid in question.
|
||||
* @param: structure containing the new RT priority.
|
||||
*/
|
||||
asmlinkage long sys_sched_setparam(pid_t pid, struct sched_param __user *param)
|
||||
SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param)
|
||||
{
|
||||
return do_sched_setscheduler(pid, -1, param);
|
||||
}
|
||||
@@ -5522,7 +5532,7 @@ asmlinkage long sys_sched_setparam(pid_t pid, struct sched_param __user *param)
|
||||
* sys_sched_getscheduler - get the policy (scheduling class) of a thread
|
||||
* @pid: the pid in question.
|
||||
*/
|
||||
asmlinkage long sys_sched_getscheduler(pid_t pid)
|
||||
SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid)
|
||||
{
|
||||
struct task_struct *p;
|
||||
int retval;
|
||||
@@ -5547,7 +5557,7 @@ asmlinkage long sys_sched_getscheduler(pid_t pid)
|
||||
* @pid: the pid in question.
|
||||
* @param: structure containing the RT priority.
|
||||
*/
|
||||
asmlinkage long sys_sched_getparam(pid_t pid, struct sched_param __user *param)
|
||||
SYSCALL_DEFINE2(sched_getparam, pid_t, pid, struct sched_param __user *, param)
|
||||
{
|
||||
struct sched_param lp;
|
||||
struct task_struct *p;
|
||||
@@ -5665,8 +5675,8 @@ static int get_user_cpu_mask(unsigned long __user *user_mask_ptr, unsigned len,
|
||||
* @len: length in bytes of the bitmask pointed to by user_mask_ptr
|
||||
* @user_mask_ptr: user-space pointer to the new cpu mask
|
||||
*/
|
||||
asmlinkage long sys_sched_setaffinity(pid_t pid, unsigned int len,
|
||||
unsigned long __user *user_mask_ptr)
|
||||
SYSCALL_DEFINE3(sched_setaffinity, pid_t, pid, unsigned int, len,
|
||||
unsigned long __user *, user_mask_ptr)
|
||||
{
|
||||
cpumask_var_t new_mask;
|
||||
int retval;
|
||||
@@ -5713,8 +5723,8 @@ out_unlock:
|
||||
* @len: length in bytes of the bitmask pointed to by user_mask_ptr
|
||||
* @user_mask_ptr: user-space pointer to hold the current cpu mask
|
||||
*/
|
||||
asmlinkage long sys_sched_getaffinity(pid_t pid, unsigned int len,
|
||||
unsigned long __user *user_mask_ptr)
|
||||
SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len,
|
||||
unsigned long __user *, user_mask_ptr)
|
||||
{
|
||||
int ret;
|
||||
cpumask_var_t mask;
|
||||
@@ -5743,7 +5753,7 @@ asmlinkage long sys_sched_getaffinity(pid_t pid, unsigned int len,
|
||||
* This function yields the current CPU to other tasks. If there are no
|
||||
* other threads running on this CPU then this function will return.
|
||||
*/
|
||||
asmlinkage long sys_sched_yield(void)
|
||||
SYSCALL_DEFINE0(sched_yield)
|
||||
{
|
||||
struct rq *rq = this_rq_lock();
|
||||
|
||||
@@ -5884,7 +5894,7 @@ long __sched io_schedule_timeout(long timeout)
|
||||
* this syscall returns the maximum rt_priority that can be used
|
||||
* by a given scheduling class.
|
||||
*/
|
||||
asmlinkage long sys_sched_get_priority_max(int policy)
|
||||
SYSCALL_DEFINE1(sched_get_priority_max, int, policy)
|
||||
{
|
||||
int ret = -EINVAL;
|
||||
|
||||
@@ -5909,7 +5919,7 @@ asmlinkage long sys_sched_get_priority_max(int policy)
|
||||
* this syscall returns the minimum rt_priority that can be used
|
||||
* by a given scheduling class.
|
||||
*/
|
||||
asmlinkage long sys_sched_get_priority_min(int policy)
|
||||
SYSCALL_DEFINE1(sched_get_priority_min, int, policy)
|
||||
{
|
||||
int ret = -EINVAL;
|
||||
|
||||
@@ -5934,8 +5944,8 @@ asmlinkage long sys_sched_get_priority_min(int policy)
|
||||
* this syscall writes the default timeslice value of a given process
|
||||
* into the user-space timespec buffer. A value of '0' means infinity.
|
||||
*/
|
||||
asmlinkage
|
||||
long sys_sched_rr_get_interval(pid_t pid, struct timespec __user *interval)
|
||||
SYSCALL_DEFINE2(sched_rr_get_interval, pid_t, pid,
|
||||
struct timespec __user *, interval)
|
||||
{
|
||||
struct task_struct *p;
|
||||
unsigned int time_slice;
|
||||
@@ -9115,6 +9125,13 @@ static int tg_schedulable(struct task_group *tg, void *data)
|
||||
runtime = d->rt_runtime;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_USER_SCHED
|
||||
if (tg == &root_task_group) {
|
||||
period = global_rt_period();
|
||||
runtime = global_rt_runtime();
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Cannot have more runtime than the period.
|
||||
*/
|
||||
|
Reference in New Issue
Block a user