Commit bdd4dfa8 authored by Ingo Molnar's avatar Ingo Molnar

sched: remove the 'u64 now' local variables

final step: remove all (now superfluous) 'u64 now' variables.

( identity transformation that causes no change in functionality. )
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent 2e1cb74a
...@@ -910,10 +910,7 @@ static int effective_prio(struct task_struct *p) ...@@ -910,10 +910,7 @@ static int effective_prio(struct task_struct *p)
*/ */
static void activate_task(struct rq *rq, struct task_struct *p, int wakeup) static void activate_task(struct rq *rq, struct task_struct *p, int wakeup)
{ {
u64 now;
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
if (p->state == TASK_UNINTERRUPTIBLE) if (p->state == TASK_UNINTERRUPTIBLE)
rq->nr_uninterruptible--; rq->nr_uninterruptible--;
...@@ -927,10 +924,7 @@ static void activate_task(struct rq *rq, struct task_struct *p, int wakeup) ...@@ -927,10 +924,7 @@ static void activate_task(struct rq *rq, struct task_struct *p, int wakeup)
*/ */
static inline void activate_idle_task(struct task_struct *p, struct rq *rq) static inline void activate_idle_task(struct task_struct *p, struct rq *rq)
{ {
u64 now;
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
if (p->state == TASK_UNINTERRUPTIBLE) if (p->state == TASK_UNINTERRUPTIBLE)
rq->nr_uninterruptible--; rq->nr_uninterruptible--;
...@@ -1647,13 +1641,11 @@ void fastcall wake_up_new_task(struct task_struct *p, unsigned long clone_flags) ...@@ -1647,13 +1641,11 @@ void fastcall wake_up_new_task(struct task_struct *p, unsigned long clone_flags)
unsigned long flags; unsigned long flags;
struct rq *rq; struct rq *rq;
int this_cpu; int this_cpu;
u64 now;
rq = task_rq_lock(p, &flags); rq = task_rq_lock(p, &flags);
BUG_ON(p->state != TASK_RUNNING); BUG_ON(p->state != TASK_RUNNING);
this_cpu = smp_processor_id(); /* parent's CPU */ this_cpu = smp_processor_id(); /* parent's CPU */
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
p->prio = effective_prio(p); p->prio = effective_prio(p);
...@@ -1955,11 +1947,9 @@ static void update_cpu_load(struct rq *this_rq) ...@@ -1955,11 +1947,9 @@ static void update_cpu_load(struct rq *this_rq)
unsigned long total_load = this_rq->ls.load.weight; unsigned long total_load = this_rq->ls.load.weight;
unsigned long this_load = total_load; unsigned long this_load = total_load;
struct load_stat *ls = &this_rq->ls; struct load_stat *ls = &this_rq->ls;
u64 now;
int i, scale; int i, scale;
__update_rq_clock(this_rq); __update_rq_clock(this_rq);
now = this_rq->clock;
this_rq->nr_load_updates++; this_rq->nr_load_updates++;
if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD))) if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD)))
...@@ -3431,7 +3421,6 @@ asmlinkage void __sched schedule(void) ...@@ -3431,7 +3421,6 @@ asmlinkage void __sched schedule(void)
struct task_struct *prev, *next; struct task_struct *prev, *next;
long *switch_count; long *switch_count;
struct rq *rq; struct rq *rq;
u64 now;
int cpu; int cpu;
need_resched: need_resched:
...@@ -3450,7 +3439,6 @@ need_resched_nonpreemptible: ...@@ -3450,7 +3439,6 @@ need_resched_nonpreemptible:
spin_lock_irq(&rq->lock); spin_lock_irq(&rq->lock);
clear_tsk_need_resched(prev); clear_tsk_need_resched(prev);
__update_rq_clock(rq); __update_rq_clock(rq);
now = rq->clock;
if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) {
if (unlikely((prev->state & TASK_INTERRUPTIBLE) && if (unlikely((prev->state & TASK_INTERRUPTIBLE) &&
...@@ -3909,13 +3897,11 @@ void rt_mutex_setprio(struct task_struct *p, int prio) ...@@ -3909,13 +3897,11 @@ void rt_mutex_setprio(struct task_struct *p, int prio)
unsigned long flags; unsigned long flags;
int oldprio, on_rq; int oldprio, on_rq;
struct rq *rq; struct rq *rq;
u64 now;
BUG_ON(prio < 0 || prio > MAX_PRIO); BUG_ON(prio < 0 || prio > MAX_PRIO);
rq = task_rq_lock(p, &flags); rq = task_rq_lock(p, &flags);
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
oldprio = p->prio; oldprio = p->prio;
on_rq = p->se.on_rq; on_rq = p->se.on_rq;
...@@ -3953,7 +3939,6 @@ void set_user_nice(struct task_struct *p, long nice) ...@@ -3953,7 +3939,6 @@ void set_user_nice(struct task_struct *p, long nice)
int old_prio, delta, on_rq; int old_prio, delta, on_rq;
unsigned long flags; unsigned long flags;
struct rq *rq; struct rq *rq;
u64 now;
if (TASK_NICE(p) == nice || nice < -20 || nice > 19) if (TASK_NICE(p) == nice || nice < -20 || nice > 19)
return; return;
...@@ -3963,7 +3948,6 @@ void set_user_nice(struct task_struct *p, long nice) ...@@ -3963,7 +3948,6 @@ void set_user_nice(struct task_struct *p, long nice)
*/ */
rq = task_rq_lock(p, &flags); rq = task_rq_lock(p, &flags);
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
/* /*
* The RT priorities are set via sched_setscheduler(), but we still * The RT priorities are set via sched_setscheduler(), but we still
* allow the 'normal' nice value to be set - but as expected * allow the 'normal' nice value to be set - but as expected
......
...@@ -667,10 +667,8 @@ static void entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) ...@@ -667,10 +667,8 @@ static void entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr)
{ {
struct rq *rq = rq_of(cfs_rq); struct rq *rq = rq_of(cfs_rq);
struct sched_entity *next; struct sched_entity *next;
u64 now;
__update_rq_clock(rq); __update_rq_clock(rq);
now = rq->clock;
/* /*
* Dequeue and enqueue the task to update its * Dequeue and enqueue the task to update its
...@@ -820,10 +818,8 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int sleep) ...@@ -820,10 +818,8 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int sleep)
static void yield_task_fair(struct rq *rq, struct task_struct *p) static void yield_task_fair(struct rq *rq, struct task_struct *p)
{ {
struct cfs_rq *cfs_rq = task_cfs_rq(p); struct cfs_rq *cfs_rq = task_cfs_rq(p);
u64 now;
__update_rq_clock(rq); __update_rq_clock(rq);
now = rq->clock;
/* /*
* Dequeue and enqueue the task to update its * Dequeue and enqueue the task to update its
* position within the tree: * position within the tree:
...@@ -1062,11 +1058,9 @@ static void set_curr_task_fair(struct rq *rq) ...@@ -1062,11 +1058,9 @@ static void set_curr_task_fair(struct rq *rq)
{ {
struct task_struct *curr = rq->curr; struct task_struct *curr = rq->curr;
struct sched_entity *se = &curr->se; struct sched_entity *se = &curr->se;
u64 now;
struct cfs_rq *cfs_rq; struct cfs_rq *cfs_rq;
update_rq_clock(rq); update_rq_clock(rq);
now = rq->clock;
for_each_sched_entity(se) { for_each_sched_entity(se) {
cfs_rq = cfs_rq_of(se); cfs_rq = cfs_rq_of(se);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment