mirror of
https://github.com/torvalds/linux.git
synced 2024-09-20 23:13:00 +00:00
tick: Use IS_ENABLED() whenever possible
Avoid ifdeferry if it can be converted to IS_ENABLED() whenever possible Signed-off-by: Frederic Weisbecker <frederic@kernel.org> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Thomas Gleixner <tglx@linutronix.de> Link: https://lore.kernel.org/r/20240225225508.11587-5-frederic@kernel.org
This commit is contained in:
parent
3aedb7fcd8
commit
27dc08096c
|
@ -111,15 +111,13 @@ void tick_handle_periodic(struct clock_event_device *dev)
|
|||
|
||||
tick_periodic(cpu);
|
||||
|
||||
#if defined(CONFIG_HIGH_RES_TIMERS) || defined(CONFIG_NO_HZ_COMMON)
|
||||
/*
|
||||
* The cpu might have transitioned to HIGHRES or NOHZ mode via
|
||||
* update_process_times() -> run_local_timers() ->
|
||||
* hrtimer_run_queues().
|
||||
*/
|
||||
if (dev->event_handler != tick_handle_periodic)
|
||||
if (IS_ENABLED(CONFIG_TICK_ONESHOT) && dev->event_handler != tick_handle_periodic)
|
||||
return;
|
||||
#endif
|
||||
|
||||
if (!clockevent_state_oneshot(dev))
|
||||
return;
|
||||
|
|
|
@ -186,7 +186,6 @@ static void tick_sched_do_timer(struct tick_sched *ts, ktime_t now)
|
|||
{
|
||||
int cpu = smp_processor_id();
|
||||
|
||||
#ifdef CONFIG_NO_HZ_COMMON
|
||||
/*
|
||||
* Check if the do_timer duty was dropped. We don't care about
|
||||
* concurrency: This happens only when the CPU in charge went
|
||||
|
@ -197,13 +196,13 @@ static void tick_sched_do_timer(struct tick_sched *ts, ktime_t now)
|
|||
* If nohz_full is enabled, this should not happen because the
|
||||
* 'tick_do_timer_cpu' CPU never relinquishes.
|
||||
*/
|
||||
if (unlikely(tick_do_timer_cpu == TICK_DO_TIMER_NONE)) {
|
||||
if (IS_ENABLED(CONFIG_NO_HZ_COMMON) &&
|
||||
unlikely(tick_do_timer_cpu == TICK_DO_TIMER_NONE)) {
|
||||
#ifdef CONFIG_NO_HZ_FULL
|
||||
WARN_ON_ONCE(tick_nohz_full_running);
|
||||
#endif
|
||||
tick_do_timer_cpu = cpu;
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Check if jiffies need an update */
|
||||
if (tick_do_timer_cpu == cpu)
|
||||
|
@ -230,7 +229,6 @@ static void tick_sched_do_timer(struct tick_sched *ts, ktime_t now)
|
|||
|
||||
static void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs)
|
||||
{
|
||||
#ifdef CONFIG_NO_HZ_COMMON
|
||||
/*
|
||||
* When we are idle and the tick is stopped, we have to touch
|
||||
* the watchdog as we might not schedule for a really long
|
||||
|
@ -239,7 +237,7 @@ static void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs)
|
|||
* idle" jiffy stamp so the idle accounting adjustment we do
|
||||
* when we go busy again does not account too many ticks.
|
||||
*/
|
||||
if (ts->tick_stopped) {
|
||||
if (IS_ENABLED(CONFIG_NO_HZ_COMMON) && ts->tick_stopped) {
|
||||
touch_softlockup_watchdog_sched();
|
||||
if (is_idle_task(current))
|
||||
ts->idle_jiffies++;
|
||||
|
@ -250,7 +248,7 @@ static void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs)
|
|||
*/
|
||||
ts->next_tick = 0;
|
||||
}
|
||||
#endif
|
||||
|
||||
update_process_times(user_mode(regs));
|
||||
profile_tick(CPU_PROFILING);
|
||||
}
|
||||
|
@ -1587,10 +1585,8 @@ void tick_cancel_sched_timer(int cpu)
|
|||
ktime_t idle_sleeptime, iowait_sleeptime;
|
||||
unsigned long idle_calls, idle_sleeps;
|
||||
|
||||
# ifdef CONFIG_HIGH_RES_TIMERS
|
||||
if (ts->sched_timer.base)
|
||||
if (IS_ENABLED(CONFIG_HIGH_RES_TIMERS) && ts->sched_timer.base)
|
||||
hrtimer_cancel(&ts->sched_timer);
|
||||
# endif
|
||||
|
||||
idle_sleeptime = ts->idle_sleeptime;
|
||||
iowait_sleeptime = ts->iowait_sleeptime;
|
||||
|
|
Loading…
Reference in New Issue
Block a user