Commit 28fbc8b6 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'sched-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip

Pull scheduler fixes from Ingo Molnar:
 "Docbook fixes that make 99% of the diffstat, plus a oneliner fix"

* 'sched-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
  sched: Ensure update_cfs_shares() is called for parents of continuously-running tasks
  sched: Fix some kernel-doc warnings
parents bfd36050 bf0bd948
...@@ -1532,6 +1532,8 @@ static inline pid_t task_pgrp_nr(struct task_struct *tsk) ...@@ -1532,6 +1532,8 @@ static inline pid_t task_pgrp_nr(struct task_struct *tsk)
* Test if a process is not yet dead (at most zombie state) * Test if a process is not yet dead (at most zombie state)
* If pid_alive fails, then pointers within the task structure * If pid_alive fails, then pointers within the task structure
* can be stale and must not be dereferenced. * can be stale and must not be dereferenced.
*
* Return: 1 if the process is alive. 0 otherwise.
*/ */
static inline int pid_alive(struct task_struct *p) static inline int pid_alive(struct task_struct *p)
{ {
...@@ -1543,6 +1545,8 @@ static inline int pid_alive(struct task_struct *p) ...@@ -1543,6 +1545,8 @@ static inline int pid_alive(struct task_struct *p)
* @tsk: Task structure to be checked. * @tsk: Task structure to be checked.
* *
* Check if a task structure is the first user space task the kernel created. * Check if a task structure is the first user space task the kernel created.
*
* Return: 1 if the task structure is init. 0 otherwise.
*/ */
static inline int is_global_init(struct task_struct *tsk) static inline int is_global_init(struct task_struct *tsk)
{ {
...@@ -1894,6 +1898,8 @@ extern struct task_struct *idle_task(int cpu); ...@@ -1894,6 +1898,8 @@ extern struct task_struct *idle_task(int cpu);
/** /**
* is_idle_task - is the specified task an idle task? * is_idle_task - is the specified task an idle task?
* @p: the task in question. * @p: the task in question.
*
* Return: 1 if @p is an idle task. 0 otherwise.
*/ */
static inline bool is_idle_task(const struct task_struct *p) static inline bool is_idle_task(const struct task_struct *p)
{ {
......
...@@ -933,6 +933,8 @@ static int effective_prio(struct task_struct *p) ...@@ -933,6 +933,8 @@ static int effective_prio(struct task_struct *p)
/** /**
* task_curr - is this task currently executing on a CPU? * task_curr - is this task currently executing on a CPU?
* @p: the task in question. * @p: the task in question.
*
* Return: 1 if the task is currently executing. 0 otherwise.
*/ */
inline int task_curr(const struct task_struct *p) inline int task_curr(const struct task_struct *p)
{ {
...@@ -1482,7 +1484,7 @@ static void ttwu_queue(struct task_struct *p, int cpu) ...@@ -1482,7 +1484,7 @@ static void ttwu_queue(struct task_struct *p, int cpu)
* the simpler "current->state = TASK_RUNNING" to mark yourself * the simpler "current->state = TASK_RUNNING" to mark yourself
* runnable without the overhead of this. * runnable without the overhead of this.
* *
* Returns %true if @p was woken up, %false if it was already running * Return: %true if @p was woken up, %false if it was already running.
* or @state didn't match @p's state. * or @state didn't match @p's state.
*/ */
static int static int
...@@ -1583,8 +1585,9 @@ static void try_to_wake_up_local(struct task_struct *p) ...@@ -1583,8 +1585,9 @@ static void try_to_wake_up_local(struct task_struct *p)
* @p: The process to be woken up. * @p: The process to be woken up.
* *
* Attempt to wake up the nominated process and move it to the set of runnable * Attempt to wake up the nominated process and move it to the set of runnable
* processes. Returns 1 if the process was woken up, 0 if it was already * processes.
* running. *
* Return: 1 if the process was woken up, 0 if it was already running.
* *
* It may be assumed that this function implies a write memory barrier before * It may be assumed that this function implies a write memory barrier before
* changing the task state if and only if any tasks are woken up. * changing the task state if and only if any tasks are woken up.
...@@ -2197,6 +2200,8 @@ void scheduler_tick(void) ...@@ -2197,6 +2200,8 @@ void scheduler_tick(void)
* This makes sure that uptime, CFS vruntime, load * This makes sure that uptime, CFS vruntime, load
* balancing, etc... continue to move forward, even * balancing, etc... continue to move forward, even
* with a very low granularity. * with a very low granularity.
*
* Return: Maximum deferment in nanoseconds.
*/ */
u64 scheduler_tick_max_deferment(void) u64 scheduler_tick_max_deferment(void)
{ {
...@@ -2808,8 +2813,8 @@ EXPORT_SYMBOL(wait_for_completion); ...@@ -2808,8 +2813,8 @@ EXPORT_SYMBOL(wait_for_completion);
* specified timeout to expire. The timeout is in jiffies. It is not * specified timeout to expire. The timeout is in jiffies. It is not
* interruptible. * interruptible.
* *
* The return value is 0 if timed out, and positive (at least 1, or number of * Return: 0 if timed out, and positive (at least 1, or number of jiffies left
* jiffies left till timeout) if completed. * till timeout) if completed.
*/ */
unsigned long __sched unsigned long __sched
wait_for_completion_timeout(struct completion *x, unsigned long timeout) wait_for_completion_timeout(struct completion *x, unsigned long timeout)
...@@ -2841,8 +2846,8 @@ EXPORT_SYMBOL(wait_for_completion_io); ...@@ -2841,8 +2846,8 @@ EXPORT_SYMBOL(wait_for_completion_io);
* specified timeout to expire. The timeout is in jiffies. It is not * specified timeout to expire. The timeout is in jiffies. It is not
* interruptible. The caller is accounted as waiting for IO. * interruptible. The caller is accounted as waiting for IO.
* *
* The return value is 0 if timed out, and positive (at least 1, or number of * Return: 0 if timed out, and positive (at least 1, or number of jiffies left
* jiffies left till timeout) if completed. * till timeout) if completed.
*/ */
unsigned long __sched unsigned long __sched
wait_for_completion_io_timeout(struct completion *x, unsigned long timeout) wait_for_completion_io_timeout(struct completion *x, unsigned long timeout)
...@@ -2858,7 +2863,7 @@ EXPORT_SYMBOL(wait_for_completion_io_timeout); ...@@ -2858,7 +2863,7 @@ EXPORT_SYMBOL(wait_for_completion_io_timeout);
* This waits for completion of a specific task to be signaled. It is * This waits for completion of a specific task to be signaled. It is
* interruptible. * interruptible.
* *
* The return value is -ERESTARTSYS if interrupted, 0 if completed. * Return: -ERESTARTSYS if interrupted, 0 if completed.
*/ */
int __sched wait_for_completion_interruptible(struct completion *x) int __sched wait_for_completion_interruptible(struct completion *x)
{ {
...@@ -2877,8 +2882,8 @@ EXPORT_SYMBOL(wait_for_completion_interruptible); ...@@ -2877,8 +2882,8 @@ EXPORT_SYMBOL(wait_for_completion_interruptible);
* This waits for either a completion of a specific task to be signaled or for a * This waits for either a completion of a specific task to be signaled or for a
* specified timeout to expire. It is interruptible. The timeout is in jiffies. * specified timeout to expire. It is interruptible. The timeout is in jiffies.
* *
* The return value is -ERESTARTSYS if interrupted, 0 if timed out, * Return: -ERESTARTSYS if interrupted, 0 if timed out, positive (at least 1,
* positive (at least 1, or number of jiffies left till timeout) if completed. * or number of jiffies left till timeout) if completed.
*/ */
long __sched long __sched
wait_for_completion_interruptible_timeout(struct completion *x, wait_for_completion_interruptible_timeout(struct completion *x,
...@@ -2895,7 +2900,7 @@ EXPORT_SYMBOL(wait_for_completion_interruptible_timeout); ...@@ -2895,7 +2900,7 @@ EXPORT_SYMBOL(wait_for_completion_interruptible_timeout);
* This waits to be signaled for completion of a specific task. It can be * This waits to be signaled for completion of a specific task. It can be
* interrupted by a kill signal. * interrupted by a kill signal.
* *
* The return value is -ERESTARTSYS if interrupted, 0 if completed. * Return: -ERESTARTSYS if interrupted, 0 if completed.
*/ */
int __sched wait_for_completion_killable(struct completion *x) int __sched wait_for_completion_killable(struct completion *x)
{ {
...@@ -2915,8 +2920,8 @@ EXPORT_SYMBOL(wait_for_completion_killable); ...@@ -2915,8 +2920,8 @@ EXPORT_SYMBOL(wait_for_completion_killable);
* signaled or for a specified timeout to expire. It can be * signaled or for a specified timeout to expire. It can be
* interrupted by a kill signal. The timeout is in jiffies. * interrupted by a kill signal. The timeout is in jiffies.
* *
* The return value is -ERESTARTSYS if interrupted, 0 if timed out, * Return: -ERESTARTSYS if interrupted, 0 if timed out, positive (at least 1,
* positive (at least 1, or number of jiffies left till timeout) if completed. * or number of jiffies left till timeout) if completed.
*/ */
long __sched long __sched
wait_for_completion_killable_timeout(struct completion *x, wait_for_completion_killable_timeout(struct completion *x,
...@@ -2930,7 +2935,7 @@ EXPORT_SYMBOL(wait_for_completion_killable_timeout); ...@@ -2930,7 +2935,7 @@ EXPORT_SYMBOL(wait_for_completion_killable_timeout);
* try_wait_for_completion - try to decrement a completion without blocking * try_wait_for_completion - try to decrement a completion without blocking
* @x: completion structure * @x: completion structure
* *
* Returns: 0 if a decrement cannot be done without blocking * Return: 0 if a decrement cannot be done without blocking
* 1 if a decrement succeeded. * 1 if a decrement succeeded.
* *
* If a completion is being used as a counting completion, * If a completion is being used as a counting completion,
...@@ -2957,7 +2962,7 @@ EXPORT_SYMBOL(try_wait_for_completion); ...@@ -2957,7 +2962,7 @@ EXPORT_SYMBOL(try_wait_for_completion);
* completion_done - Test to see if a completion has any waiters * completion_done - Test to see if a completion has any waiters
* @x: completion structure * @x: completion structure
* *
* Returns: 0 if there are waiters (wait_for_completion() in progress) * Return: 0 if there are waiters (wait_for_completion() in progress)
* 1 if there are no waiters. * 1 if there are no waiters.
* *
*/ */
...@@ -3194,7 +3199,7 @@ SYSCALL_DEFINE1(nice, int, increment) ...@@ -3194,7 +3199,7 @@ SYSCALL_DEFINE1(nice, int, increment)
* task_prio - return the priority value of a given task. * task_prio - return the priority value of a given task.
* @p: the task in question. * @p: the task in question.
* *
* This is the priority value as seen by users in /proc. * Return: The priority value as seen by users in /proc.
* RT tasks are offset by -200. Normal tasks are centered * RT tasks are offset by -200. Normal tasks are centered
* around 0, value goes from -16 to +15. * around 0, value goes from -16 to +15.
*/ */
...@@ -3206,6 +3211,8 @@ int task_prio(const struct task_struct *p) ...@@ -3206,6 +3211,8 @@ int task_prio(const struct task_struct *p)
/** /**
* task_nice - return the nice value of a given task. * task_nice - return the nice value of a given task.
* @p: the task in question. * @p: the task in question.
*
* Return: The nice value [ -20 ... 0 ... 19 ].
*/ */
int task_nice(const struct task_struct *p) int task_nice(const struct task_struct *p)
{ {
...@@ -3216,6 +3223,8 @@ EXPORT_SYMBOL(task_nice); ...@@ -3216,6 +3223,8 @@ EXPORT_SYMBOL(task_nice);
/** /**
* idle_cpu - is a given cpu idle currently? * idle_cpu - is a given cpu idle currently?
* @cpu: the processor in question. * @cpu: the processor in question.
*
* Return: 1 if the CPU is currently idle. 0 otherwise.
*/ */
int idle_cpu(int cpu) int idle_cpu(int cpu)
{ {
...@@ -3238,6 +3247,8 @@ int idle_cpu(int cpu) ...@@ -3238,6 +3247,8 @@ int idle_cpu(int cpu)
/** /**
* idle_task - return the idle task for a given cpu. * idle_task - return the idle task for a given cpu.
* @cpu: the processor in question. * @cpu: the processor in question.
*
* Return: The idle task for the cpu @cpu.
*/ */
struct task_struct *idle_task(int cpu) struct task_struct *idle_task(int cpu)
{ {
...@@ -3247,6 +3258,8 @@ struct task_struct *idle_task(int cpu) ...@@ -3247,6 +3258,8 @@ struct task_struct *idle_task(int cpu)
/** /**
* find_process_by_pid - find a process with a matching PID value. * find_process_by_pid - find a process with a matching PID value.
* @pid: the pid in question. * @pid: the pid in question.
*
* The task of @pid, if found. %NULL otherwise.
*/ */
static struct task_struct *find_process_by_pid(pid_t pid) static struct task_struct *find_process_by_pid(pid_t pid)
{ {
...@@ -3444,6 +3457,8 @@ static int __sched_setscheduler(struct task_struct *p, int policy, ...@@ -3444,6 +3457,8 @@ static int __sched_setscheduler(struct task_struct *p, int policy,
* @policy: new policy. * @policy: new policy.
* @param: structure containing the new RT priority. * @param: structure containing the new RT priority.
* *
* Return: 0 on success. An error code otherwise.
*
* NOTE that the task may be already dead. * NOTE that the task may be already dead.
*/ */
int sched_setscheduler(struct task_struct *p, int policy, int sched_setscheduler(struct task_struct *p, int policy,
...@@ -3463,6 +3478,8 @@ EXPORT_SYMBOL_GPL(sched_setscheduler); ...@@ -3463,6 +3478,8 @@ EXPORT_SYMBOL_GPL(sched_setscheduler);
* current context has permission. For example, this is needed in * current context has permission. For example, this is needed in
* stop_machine(): we create temporary high priority worker threads, * stop_machine(): we create temporary high priority worker threads,
* but our caller might not have that capability. * but our caller might not have that capability.
*
* Return: 0 on success. An error code otherwise.
*/ */
int sched_setscheduler_nocheck(struct task_struct *p, int policy, int sched_setscheduler_nocheck(struct task_struct *p, int policy,
const struct sched_param *param) const struct sched_param *param)
...@@ -3497,6 +3514,8 @@ do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param) ...@@ -3497,6 +3514,8 @@ do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
* @pid: the pid in question. * @pid: the pid in question.
* @policy: new policy. * @policy: new policy.
* @param: structure containing the new RT priority. * @param: structure containing the new RT priority.
*
* Return: 0 on success. An error code otherwise.
*/ */
SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy, SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy,
struct sched_param __user *, param) struct sched_param __user *, param)
...@@ -3512,6 +3531,8 @@ SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy, ...@@ -3512,6 +3531,8 @@ SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy,
* sys_sched_setparam - set/change the RT priority of a thread * sys_sched_setparam - set/change the RT priority of a thread
* @pid: the pid in question. * @pid: the pid in question.
* @param: structure containing the new RT priority. * @param: structure containing the new RT priority.
*
* Return: 0 on success. An error code otherwise.
*/ */
SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param) SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param)
{ {
...@@ -3521,6 +3542,9 @@ SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param) ...@@ -3521,6 +3542,9 @@ SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param)
/** /**
* sys_sched_getscheduler - get the policy (scheduling class) of a thread * sys_sched_getscheduler - get the policy (scheduling class) of a thread
* @pid: the pid in question. * @pid: the pid in question.
*
* Return: On success, the policy of the thread. Otherwise, a negative error
* code.
*/ */
SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid) SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid)
{ {
...@@ -3547,6 +3571,9 @@ SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid) ...@@ -3547,6 +3571,9 @@ SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid)
* sys_sched_getparam - get the RT priority of a thread * sys_sched_getparam - get the RT priority of a thread
* @pid: the pid in question. * @pid: the pid in question.
* @param: structure containing the RT priority. * @param: structure containing the RT priority.
*
* Return: On success, 0 and the RT priority is in @param. Otherwise, an error
* code.
*/ */
SYSCALL_DEFINE2(sched_getparam, pid_t, pid, struct sched_param __user *, param) SYSCALL_DEFINE2(sched_getparam, pid_t, pid, struct sched_param __user *, param)
{ {
...@@ -3671,6 +3698,8 @@ static int get_user_cpu_mask(unsigned long __user *user_mask_ptr, unsigned len, ...@@ -3671,6 +3698,8 @@ static int get_user_cpu_mask(unsigned long __user *user_mask_ptr, unsigned len,
* @pid: pid of the process * @pid: pid of the process
* @len: length in bytes of the bitmask pointed to by user_mask_ptr * @len: length in bytes of the bitmask pointed to by user_mask_ptr
* @user_mask_ptr: user-space pointer to the new cpu mask * @user_mask_ptr: user-space pointer to the new cpu mask
*
* Return: 0 on success. An error code otherwise.
*/ */
SYSCALL_DEFINE3(sched_setaffinity, pid_t, pid, unsigned int, len, SYSCALL_DEFINE3(sched_setaffinity, pid_t, pid, unsigned int, len,
unsigned long __user *, user_mask_ptr) unsigned long __user *, user_mask_ptr)
...@@ -3722,6 +3751,8 @@ long sched_getaffinity(pid_t pid, struct cpumask *mask) ...@@ -3722,6 +3751,8 @@ long sched_getaffinity(pid_t pid, struct cpumask *mask)
* @pid: pid of the process * @pid: pid of the process
* @len: length in bytes of the bitmask pointed to by user_mask_ptr * @len: length in bytes of the bitmask pointed to by user_mask_ptr
* @user_mask_ptr: user-space pointer to hold the current cpu mask * @user_mask_ptr: user-space pointer to hold the current cpu mask
*
* Return: 0 on success. An error code otherwise.
*/ */
SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len, SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len,
unsigned long __user *, user_mask_ptr) unsigned long __user *, user_mask_ptr)
...@@ -3756,6 +3787,8 @@ SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len, ...@@ -3756,6 +3787,8 @@ SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len,
* *
* This function yields the current CPU to other tasks. If there are no * This function yields the current CPU to other tasks. If there are no
* other threads running on this CPU then this function will return. * other threads running on this CPU then this function will return.
*
* Return: 0.
*/ */
SYSCALL_DEFINE0(sched_yield) SYSCALL_DEFINE0(sched_yield)
{ {
...@@ -3881,7 +3914,7 @@ EXPORT_SYMBOL(yield); ...@@ -3881,7 +3914,7 @@ EXPORT_SYMBOL(yield);
* It's the caller's job to ensure that the target task struct * It's the caller's job to ensure that the target task struct
* can't go away on us before we can do any checks. * can't go away on us before we can do any checks.
* *
* Returns: * Return:
* true (>0) if we indeed boosted the target task. * true (>0) if we indeed boosted the target task.
* false (0) if we failed to boost the target. * false (0) if we failed to boost the target.
* -ESRCH if there's no task to yield to. * -ESRCH if there's no task to yield to.
...@@ -3984,8 +4017,9 @@ long __sched io_schedule_timeout(long timeout) ...@@ -3984,8 +4017,9 @@ long __sched io_schedule_timeout(long timeout)
* sys_sched_get_priority_max - return maximum RT priority. * sys_sched_get_priority_max - return maximum RT priority.
* @policy: scheduling class. * @policy: scheduling class.
* *
* this syscall returns the maximum rt_priority that can be used * Return: On success, this syscall returns the maximum
* by a given scheduling class. * rt_priority that can be used by a given scheduling class.
* On failure, a negative error code is returned.
*/ */
SYSCALL_DEFINE1(sched_get_priority_max, int, policy) SYSCALL_DEFINE1(sched_get_priority_max, int, policy)
{ {
...@@ -4009,8 +4043,9 @@ SYSCALL_DEFINE1(sched_get_priority_max, int, policy) ...@@ -4009,8 +4043,9 @@ SYSCALL_DEFINE1(sched_get_priority_max, int, policy)
* sys_sched_get_priority_min - return minimum RT priority. * sys_sched_get_priority_min - return minimum RT priority.
* @policy: scheduling class. * @policy: scheduling class.
* *
* this syscall returns the minimum rt_priority that can be used * Return: On success, this syscall returns the minimum
* by a given scheduling class. * rt_priority that can be used by a given scheduling class.
* On failure, a negative error code is returned.
*/ */
SYSCALL_DEFINE1(sched_get_priority_min, int, policy) SYSCALL_DEFINE1(sched_get_priority_min, int, policy)
{ {
...@@ -4036,6 +4071,9 @@ SYSCALL_DEFINE1(sched_get_priority_min, int, policy) ...@@ -4036,6 +4071,9 @@ SYSCALL_DEFINE1(sched_get_priority_min, int, policy)
* *
* this syscall writes the default timeslice value of a given process * this syscall writes the default timeslice value of a given process
* into the user-space timespec buffer. A value of '0' means infinity. * into the user-space timespec buffer. A value of '0' means infinity.
*
* Return: On success, 0 and the timeslice is in @interval. Otherwise,
* an error code.
*/ */
SYSCALL_DEFINE2(sched_rr_get_interval, pid_t, pid, SYSCALL_DEFINE2(sched_rr_get_interval, pid_t, pid,
struct timespec __user *, interval) struct timespec __user *, interval)
...@@ -6644,6 +6682,8 @@ void normalize_rt_tasks(void) ...@@ -6644,6 +6682,8 @@ void normalize_rt_tasks(void)
* @cpu: the processor in question. * @cpu: the processor in question.
* *
* ONLY VALID WHEN THE WHOLE SYSTEM IS STOPPED! * ONLY VALID WHEN THE WHOLE SYSTEM IS STOPPED!
*
* Return: The current task for @cpu.
*/ */
struct task_struct *curr_task(int cpu) struct task_struct *curr_task(int cpu)
{ {
......
...@@ -62,7 +62,7 @@ static int convert_prio(int prio) ...@@ -62,7 +62,7 @@ static int convert_prio(int prio)
* any discrepancies created by racing against the uncertainty of the current * any discrepancies created by racing against the uncertainty of the current
* priority configuration. * priority configuration.
* *
* Returns: (int)bool - CPUs were found * Return: (int)bool - CPUs were found
*/ */
int cpupri_find(struct cpupri *cp, struct task_struct *p, int cpupri_find(struct cpupri *cp, struct task_struct *p,
struct cpumask *lowest_mask) struct cpumask *lowest_mask)
...@@ -203,7 +203,7 @@ void cpupri_set(struct cpupri *cp, int cpu, int newpri) ...@@ -203,7 +203,7 @@ void cpupri_set(struct cpupri *cp, int cpu, int newpri)
* cpupri_init - initialize the cpupri structure * cpupri_init - initialize the cpupri structure
* @cp: The cpupri context * @cp: The cpupri context
* *
* Returns: -ENOMEM if memory fails. * Return: -ENOMEM on memory allocation failure.
*/ */
int cpupri_init(struct cpupri *cp) int cpupri_init(struct cpupri *cp)
{ {
......
...@@ -2032,6 +2032,7 @@ entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) ...@@ -2032,6 +2032,7 @@ entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued)
*/ */
update_entity_load_avg(curr, 1); update_entity_load_avg(curr, 1);
update_cfs_rq_blocked_load(cfs_rq, 1); update_cfs_rq_blocked_load(cfs_rq, 1);
update_cfs_shares(cfs_rq);
#ifdef CONFIG_SCHED_HRTICK #ifdef CONFIG_SCHED_HRTICK
/* /*
...@@ -4280,6 +4281,8 @@ struct sg_lb_stats { ...@@ -4280,6 +4281,8 @@ struct sg_lb_stats {
* get_sd_load_idx - Obtain the load index for a given sched domain. * get_sd_load_idx - Obtain the load index for a given sched domain.
* @sd: The sched_domain whose load_idx is to be obtained. * @sd: The sched_domain whose load_idx is to be obtained.
* @idle: The Idle status of the CPU for whose sd load_icx is obtained. * @idle: The Idle status of the CPU for whose sd load_icx is obtained.
*
* Return: The load index.
*/ */
static inline int get_sd_load_idx(struct sched_domain *sd, static inline int get_sd_load_idx(struct sched_domain *sd,
enum cpu_idle_type idle) enum cpu_idle_type idle)
...@@ -4574,6 +4577,9 @@ static inline void update_sg_lb_stats(struct lb_env *env, ...@@ -4574,6 +4577,9 @@ static inline void update_sg_lb_stats(struct lb_env *env,
* *
* Determine if @sg is a busier group than the previously selected * Determine if @sg is a busier group than the previously selected
* busiest group. * busiest group.
*
* Return: %true if @sg is a busier group than the previously selected
* busiest group. %false otherwise.
*/ */
static bool update_sd_pick_busiest(struct lb_env *env, static bool update_sd_pick_busiest(struct lb_env *env,
struct sd_lb_stats *sds, struct sd_lb_stats *sds,
...@@ -4691,7 +4697,7 @@ static inline void update_sd_lb_stats(struct lb_env *env, ...@@ -4691,7 +4697,7 @@ static inline void update_sd_lb_stats(struct lb_env *env,
* assuming lower CPU number will be equivalent to lower a SMT thread * assuming lower CPU number will be equivalent to lower a SMT thread
* number. * number.
* *
* Returns 1 when packing is required and a task should be moved to * Return: 1 when packing is required and a task should be moved to
* this CPU. The amount of the imbalance is returned in *imbalance. * this CPU. The amount of the imbalance is returned in *imbalance.
* *
* @env: The load balancing environment. * @env: The load balancing environment.
...@@ -4869,7 +4875,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s ...@@ -4869,7 +4875,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s
* @balance: Pointer to a variable indicating if this_cpu * @balance: Pointer to a variable indicating if this_cpu
* is the appropriate cpu to perform load balancing at this_level. * is the appropriate cpu to perform load balancing at this_level.
* *
* Returns: - the busiest group if imbalance exists. * Return: - The busiest group if imbalance exists.
* - If no imbalance and user has opted for power-savings balance, * - If no imbalance and user has opted for power-savings balance,
* return the least loaded group whose CPUs can be * return the least loaded group whose CPUs can be
* put to idle by rebalancing its tasks onto our group. * put to idle by rebalancing its tasks onto our group.
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment