2012-05-21 19:50:07 -07:00
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
|
2008-03-10 15:28:04 -07:00
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/smp.h>
|
2009-02-27 13:25:28 -08:00
|
|
|
#include <linux/prctl.h>
|
2008-03-10 15:28:04 -07:00
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/sched.h>
|
2016-07-13 20:18:56 -04:00
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/export.h>
|
2008-04-25 17:39:01 +02:00
|
|
|
#include <linux/pm.h>
|
2015-04-03 02:01:28 +02:00
|
|
|
#include <linux/tick.h>
|
2009-05-11 22:05:28 -04:00
|
|
|
#include <linux/random.h>
|
2009-09-19 09:40:22 +03:00
|
|
|
#include <linux/user-return-notifier.h>
|
2009-12-08 00:29:42 -08:00
|
|
|
#include <linux/dmi.h>
|
|
|
|
#include <linux/utsname.h>
|
2012-03-25 23:00:04 +02:00
|
|
|
#include <linux/stackprotector.h>
|
|
|
|
#include <linux/tick.h>
|
|
|
|
#include <linux/cpuidle.h>
|
2009-09-17 16:11:28 +02:00
|
|
|
#include <trace/events/power.h>
|
2009-09-09 19:22:48 +02:00
|
|
|
#include <linux/hw_breakpoint.h>
|
2011-01-20 15:42:52 +01:00
|
|
|
#include <asm/cpu.h>
|
2008-11-11 14:33:44 +01:00
|
|
|
#include <asm/apic.h>
|
2009-04-11 00:03:10 +05:30
|
|
|
#include <asm/syscalls.h>
|
2009-02-27 13:25:28 -08:00
|
|
|
#include <asm/idle.h>
|
|
|
|
#include <asm/uaccess.h>
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
#include <asm/mwait.h>
|
2015-04-24 02:54:44 +02:00
|
|
|
#include <asm/fpu/internal.h>
|
2009-06-01 23:44:55 +05:30
|
|
|
#include <asm/debugreg.h>
|
2012-03-25 23:00:04 +02:00
|
|
|
#include <asm/nmi.h>
|
2014-10-24 15:58:07 -07:00
|
|
|
#include <asm/tlbflush.h>
|
2015-08-12 18:29:40 +02:00
|
|
|
#include <asm/mce.h>
|
2015-07-29 01:41:16 -04:00
|
|
|
#include <asm/vm86.h>
|
2016-08-13 12:38:18 -04:00
|
|
|
#include <asm/switch_to.h>
|
2018-04-29 15:21:42 +02:00
|
|
|
#include <asm/spec-ctrl.h>
|
2012-03-25 23:00:04 +02:00
|
|
|
|
2018-11-25 19:33:47 +01:00
|
|
|
#include "process.h"
|
|
|
|
|
2012-05-03 09:03:01 +00:00
|
|
|
/*
|
|
|
|
* per-CPU TSS segments. Threads are completely 'soft' on Linux,
|
|
|
|
* no more per-task TSS's. The TSS size is kept cacheline-aligned
|
|
|
|
* so they are allowed to end up in the .data..cacheline_aligned
|
|
|
|
* section. Since TSS's are completely CPU-local, we want them
|
|
|
|
* on exact cacheline boundaries, to eliminate cacheline ping-pong.
|
|
|
|
*/
|
2017-05-04 14:26:50 +02:00
|
|
|
__visible DEFINE_PER_CPU_SHARED_ALIGNED_USER_MAPPED(struct tss_struct, cpu_tss) = {
|
2015-03-05 19:19:06 -08:00
|
|
|
.x86_tss = {
|
2015-03-10 11:05:59 -07:00
|
|
|
.sp0 = TOP_OF_INIT_STACK,
|
2015-03-05 19:19:06 -08:00
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
.ss0 = __KERNEL_DS,
|
|
|
|
.ss1 = __KERNEL_CS,
|
|
|
|
.io_bitmap_base = INVALID_IO_BITMAP_OFFSET,
|
|
|
|
#endif
|
|
|
|
},
|
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
/*
|
|
|
|
* Note that the .io_bitmap member must be extra-big. This is because
|
|
|
|
* the CPU will access an additional byte beyond the end of the IO
|
|
|
|
* permission bitmap. The extra byte must be all 1 bits, and must
|
|
|
|
* be within the limit.
|
|
|
|
*/
|
|
|
|
.io_bitmap = { [0 ... IO_BITMAP_LONGS] = ~0 },
|
|
|
|
#endif
|
2016-03-09 19:00:33 -08:00
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
.SYSENTER_stack_canary = STACK_END_MAGIC,
|
|
|
|
#endif
|
2015-03-05 19:19:06 -08:00
|
|
|
};
|
2015-05-04 15:16:44 -03:00
|
|
|
EXPORT_PER_CPU_SYMBOL(cpu_tss);
|
2012-05-03 09:03:01 +00:00
|
|
|
|
2012-03-25 23:00:04 +02:00
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
static DEFINE_PER_CPU(unsigned char, is_idle);
|
|
|
|
#endif
|
2008-06-24 17:58:53 +08:00
|
|
|
|
2012-05-16 15:03:51 -07:00
|
|
|
/*
|
|
|
|
* this gets called so that we can store lazy state into memory and copy the
|
|
|
|
* current task into the new thread.
|
|
|
|
*/
|
2008-03-10 15:28:04 -07:00
|
|
|
int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src)
|
|
|
|
{
|
2015-07-17 12:28:12 +02:00
|
|
|
memcpy(dst, src, arch_task_struct_size);
|
2015-10-30 22:42:46 -07:00
|
|
|
#ifdef CONFIG_VM86
|
|
|
|
dst->thread.vm86 = NULL;
|
|
|
|
#endif
|
2014-09-02 19:57:23 +02:00
|
|
|
|
2015-04-24 02:07:15 +02:00
|
|
|
return fpu__copy(&dst->thread.fpu, &src->thread.fpu);
|
2008-03-10 15:28:04 -07:00
|
|
|
}
|
2008-04-25 17:39:01 +02:00
|
|
|
|
2009-02-27 13:25:28 -08:00
|
|
|
/*
|
|
|
|
* Free current thread data structures etc..
|
|
|
|
*/
|
2016-05-20 17:00:20 -07:00
|
|
|
void exit_thread(struct task_struct *tsk)
|
2009-02-27 13:25:28 -08:00
|
|
|
{
|
2016-05-20 17:00:20 -07:00
|
|
|
struct thread_struct *t = &tsk->thread;
|
2009-03-16 13:07:21 +01:00
|
|
|
unsigned long *bp = t->io_bitmap_ptr;
|
2015-04-23 12:33:50 +02:00
|
|
|
struct fpu *fpu = &t->fpu;
|
2009-02-27 13:25:28 -08:00
|
|
|
|
2009-03-16 13:07:21 +01:00
|
|
|
if (bp) {
|
2015-03-05 19:19:05 -08:00
|
|
|
struct tss_struct *tss = &per_cpu(cpu_tss, get_cpu());
|
2009-02-27 13:25:28 -08:00
|
|
|
|
|
|
|
t->io_bitmap_ptr = NULL;
|
|
|
|
clear_thread_flag(TIF_IO_BITMAP);
|
|
|
|
/*
|
|
|
|
* Careful, clear this in the TSS too:
|
|
|
|
*/
|
|
|
|
memset(tss->io_bitmap, 0xff, t->io_bitmap_max);
|
|
|
|
t->io_bitmap_max = 0;
|
|
|
|
put_cpu();
|
2009-03-16 13:07:21 +01:00
|
|
|
kfree(bp);
|
2009-02-27 13:25:28 -08:00
|
|
|
}
|
2012-05-16 15:03:54 -07:00
|
|
|
|
2015-07-29 01:41:16 -04:00
|
|
|
free_vm86(t);
|
|
|
|
|
x86/fpu: Synchronize the naming of drop_fpu() and fpu_reset_state()
drop_fpu() and fpu_reset_state() are similar in functionality
and in scope, yet this is not apparent from their names.
drop_fpu() deactivates FPU contents (both the fpregs and the fpstate),
but leaves register contents intact in the eager-FPU case, mostly as an
optimization. It disables fpregs in the lazy FPU case. The drop_fpu()
method can be used to destroy FPU state in an optimized way, when we
know that a new state will be loaded before user-space might see
any remains of the old FPU state:
- such as in sys_exit()'s exit_thread() where we know this task
won't execute any user-space instructions anymore and the
next context switch cleans up the FPU. The old FPU state
might still be around in the eagerfpu case but won't be
saved.
- in __restore_xstate_sig(), where we use drop_fpu() before
copying a new state into the fpstate and activating that one.
No user-pace instructions can execute between those steps.
- in sys_execve()'s fpu__clear(): there we use drop_fpu() in
the !eagerfpu case, where it's equivalent to a full reinit.
fpu_reset_state() is a stronger version of drop_fpu(): both in
the eagerfpu and the lazy-FPU case it guarantees that fpregs
are reinitialized to init state. This method is used in cases
where we need a full reset:
- handle_signal() uses fpu_reset_state() to reset the FPU state
to init before executing a user-space signal handler. While we
have already saved the original FPU state at this point, and
always restore the original state, the signal handling code
still has to do this reinit, because signals may interrupt
any user-space instruction, and the FPU might be in various
intermediate states (such as an unbalanced x87 stack) that is
not immediately usable for general C signal handler code.
- __restore_xstate_sig() uses fpu_reset_state() when the signal
frame has no FP context. Since the signal handler may have
modified the FPU state, it gets reset back to init state.
- in another branch __restore_xstate_sig() uses fpu_reset_state()
to handle a restoration error: when restore_user_xstate() fails
to restore FPU state and we might have inconsistent FPU data,
fpu_reset_state() is used to reset it back to a known good
state.
- __kernel_fpu_end() uses fpu_reset_state() in an error branch.
This is in a 'must not trigger' error branch, so on bug-free
kernels this never triggers.
- fpu__restore() uses fpu_reset_state() in an error path
as well: if the fpstate was set up with invalid FPU state
(via ptrace or via a signal handler), then it's reset back
to init state.
- likewise, the scheduler's switch_fpu_finish() uses it in a
restoration error path too.
Move both drop_fpu() and fpu_reset_state() to the fpu__*() namespace
and harmonize their naming with their function:
fpu__drop()
fpu__reset()
This clearly shows that both methods operate on the full state of the
FPU, just like fpu__restore().
Also add comments to explain what each function does.
Cc: Andy Lutomirski <luto@amacapital.net>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Fenghua Yu <fenghua.yu@intel.com>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Oleg Nesterov <oleg@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2015-04-29 19:04:31 +02:00
|
|
|
fpu__drop(fpu);
|
2009-02-27 13:25:28 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
void flush_thread(void)
|
|
|
|
{
|
|
|
|
struct task_struct *tsk = current;
|
|
|
|
|
2009-09-09 19:22:48 +02:00
|
|
|
flush_ptrace_hw_breakpoint(tsk);
|
2009-02-27 13:25:28 -08:00
|
|
|
memset(tsk->thread.tls_array, 0, sizeof(tsk->thread.tls_array));
|
2015-01-19 19:52:12 +01:00
|
|
|
|
2015-04-29 20:35:33 +02:00
|
|
|
fpu__clear(&tsk->thread.fpu);
|
2009-02-27 13:25:28 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
void disable_TSC(void)
|
|
|
|
{
|
|
|
|
preempt_disable();
|
|
|
|
if (!test_and_set_thread_flag(TIF_NOTSC))
|
|
|
|
/*
|
|
|
|
* Must flip the CPU state synchronously with
|
|
|
|
* TIF_NOTSC in the current running context.
|
|
|
|
*/
|
2017-02-14 00:11:04 -08:00
|
|
|
cr4_set_bits(X86_CR4_TSD);
|
2009-02-27 13:25:28 -08:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
static void enable_TSC(void)
|
|
|
|
{
|
|
|
|
preempt_disable();
|
|
|
|
if (test_and_clear_thread_flag(TIF_NOTSC))
|
|
|
|
/*
|
|
|
|
* Must flip the CPU state synchronously with
|
|
|
|
* TIF_NOTSC in the current running context.
|
|
|
|
*/
|
2017-02-14 00:11:04 -08:00
|
|
|
cr4_clear_bits(X86_CR4_TSD);
|
2009-02-27 13:25:28 -08:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
int get_tsc_mode(unsigned long adr)
|
|
|
|
{
|
|
|
|
unsigned int val;
|
|
|
|
|
|
|
|
if (test_thread_flag(TIF_NOTSC))
|
|
|
|
val = PR_TSC_SIGSEGV;
|
|
|
|
else
|
|
|
|
val = PR_TSC_ENABLE;
|
|
|
|
|
|
|
|
return put_user(val, (unsigned int __user *)adr);
|
|
|
|
}
|
|
|
|
|
|
|
|
int set_tsc_mode(unsigned int val)
|
|
|
|
{
|
|
|
|
if (val == PR_TSC_SIGSEGV)
|
|
|
|
disable_TSC();
|
|
|
|
else if (val == PR_TSC_ENABLE)
|
|
|
|
enable_TSC();
|
|
|
|
else
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-11-25 19:33:47 +01:00
|
|
|
static inline void switch_to_bitmap(struct thread_struct *prev,
|
2017-02-14 00:11:02 -08:00
|
|
|
struct thread_struct *next,
|
|
|
|
unsigned long tifp, unsigned long tifn)
|
2009-02-27 13:25:28 -08:00
|
|
|
{
|
2018-11-25 19:33:47 +01:00
|
|
|
struct tss_struct *tss = this_cpu_ptr(&cpu_tss);
|
|
|
|
|
2017-02-14 00:11:02 -08:00
|
|
|
if (tifn & _TIF_IO_BITMAP) {
|
2009-02-27 13:25:28 -08:00
|
|
|
/*
|
|
|
|
* Copy the relevant range of the IO bitmap.
|
|
|
|
* Normally this is 128 bytes or less:
|
|
|
|
*/
|
|
|
|
memcpy(tss->io_bitmap, next->io_bitmap_ptr,
|
|
|
|
max(prev->io_bitmap_max, next->io_bitmap_max));
|
2017-02-14 00:11:02 -08:00
|
|
|
} else if (tifp & _TIF_IO_BITMAP) {
|
2009-02-27 13:25:28 -08:00
|
|
|
/*
|
|
|
|
* Clear any possible leftover bits:
|
|
|
|
*/
|
|
|
|
memset(tss->io_bitmap, 0xff, prev->io_bitmap_max);
|
|
|
|
}
|
2017-02-14 00:11:02 -08:00
|
|
|
}
|
|
|
|
|
2018-05-09 21:53:09 +02:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
|
|
|
|
struct ssb_state {
|
|
|
|
struct ssb_state *shared_state;
|
|
|
|
raw_spinlock_t lock;
|
|
|
|
unsigned int disable_state;
|
|
|
|
unsigned long local_state;
|
|
|
|
};
|
|
|
|
|
|
|
|
#define LSTATE_SSB 0
|
|
|
|
|
|
|
|
static DEFINE_PER_CPU(struct ssb_state, ssb_state);
|
|
|
|
|
|
|
|
void speculative_store_bypass_ht_init(void)
|
2018-04-29 15:21:42 +02:00
|
|
|
{
|
2018-05-09 21:53:09 +02:00
|
|
|
struct ssb_state *st = this_cpu_ptr(&ssb_state);
|
|
|
|
unsigned int this_cpu = smp_processor_id();
|
|
|
|
unsigned int cpu;
|
|
|
|
|
|
|
|
st->local_state = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Shared state setup happens once on the first bringup
|
|
|
|
* of the CPU. It's not destroyed on CPU hotunplug.
|
|
|
|
*/
|
|
|
|
if (st->shared_state)
|
|
|
|
return;
|
|
|
|
|
|
|
|
raw_spin_lock_init(&st->lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Go over HT siblings and check whether one of them has set up the
|
|
|
|
* shared state pointer already.
|
|
|
|
*/
|
|
|
|
for_each_cpu(cpu, topology_sibling_cpumask(this_cpu)) {
|
|
|
|
if (cpu == this_cpu)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!per_cpu(ssb_state, cpu).shared_state)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Link it to the state of the sibling: */
|
|
|
|
st->shared_state = per_cpu(ssb_state, cpu).shared_state;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* First HT sibling to come up on the core. Link shared state of
|
|
|
|
* the first HT sibling to itself. The siblings on the same core
|
|
|
|
* which come up later will see the shared state pointer and link
|
|
|
|
* themself to the state of this CPU.
|
|
|
|
*/
|
|
|
|
st->shared_state = st;
|
|
|
|
}
|
2018-04-29 15:21:42 +02:00
|
|
|
|
2018-05-09 21:53:09 +02:00
|
|
|
/*
|
|
|
|
* Logic is: First HT sibling enables SSBD for both siblings in the core
|
|
|
|
* and last sibling to disable it, disables it for the whole core. This how
|
|
|
|
* MSR_SPEC_CTRL works in "hardware":
|
|
|
|
*
|
|
|
|
* CORE_SPEC_CTRL = THREAD0_SPEC_CTRL | THREAD1_SPEC_CTRL
|
|
|
|
*/
|
|
|
|
static __always_inline void amd_set_core_ssb_state(unsigned long tifn)
|
|
|
|
{
|
|
|
|
struct ssb_state *st = this_cpu_ptr(&ssb_state);
|
|
|
|
u64 msr = x86_amd_ls_cfg_base;
|
|
|
|
|
|
|
|
if (!static_cpu_has(X86_FEATURE_ZEN)) {
|
|
|
|
msr |= ssbd_tif_to_amd_ls_cfg(tifn);
|
2018-04-29 15:21:42 +02:00
|
|
|
wrmsrl(MSR_AMD64_LS_CFG, msr);
|
2018-05-09 21:53:09 +02:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tifn & _TIF_SSBD) {
|
|
|
|
/*
|
|
|
|
* Since this can race with prctl(), block reentry on the
|
|
|
|
* same CPU.
|
|
|
|
*/
|
|
|
|
if (__test_and_set_bit(LSTATE_SSB, &st->local_state))
|
|
|
|
return;
|
|
|
|
|
|
|
|
msr |= x86_amd_ls_cfg_ssbd_mask;
|
|
|
|
|
|
|
|
raw_spin_lock(&st->shared_state->lock);
|
|
|
|
/* First sibling enables SSBD: */
|
|
|
|
if (!st->shared_state->disable_state)
|
|
|
|
wrmsrl(MSR_AMD64_LS_CFG, msr);
|
|
|
|
st->shared_state->disable_state++;
|
|
|
|
raw_spin_unlock(&st->shared_state->lock);
|
2018-04-29 15:21:42 +02:00
|
|
|
} else {
|
2018-05-09 21:53:09 +02:00
|
|
|
if (!__test_and_clear_bit(LSTATE_SSB, &st->local_state))
|
|
|
|
return;
|
|
|
|
|
|
|
|
raw_spin_lock(&st->shared_state->lock);
|
|
|
|
st->shared_state->disable_state--;
|
|
|
|
if (!st->shared_state->disable_state)
|
|
|
|
wrmsrl(MSR_AMD64_LS_CFG, msr);
|
|
|
|
raw_spin_unlock(&st->shared_state->lock);
|
2018-04-29 15:21:42 +02:00
|
|
|
}
|
|
|
|
}
|
2018-05-09 21:53:09 +02:00
|
|
|
#else
|
|
|
|
static __always_inline void amd_set_core_ssb_state(unsigned long tifn)
|
|
|
|
{
|
|
|
|
u64 msr = x86_amd_ls_cfg_base | ssbd_tif_to_amd_ls_cfg(tifn);
|
|
|
|
|
|
|
|
wrmsrl(MSR_AMD64_LS_CFG, msr);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2018-05-17 17:09:18 +02:00
|
|
|
static __always_inline void amd_set_ssb_virt_state(unsigned long tifn)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* SSBD has the same definition in SPEC_CTRL and VIRT_SPEC_CTRL,
|
|
|
|
* so ssbd_tif_to_spec_ctrl() just works.
|
|
|
|
*/
|
|
|
|
wrmsrl(MSR_AMD64_VIRT_SPEC_CTRL, ssbd_tif_to_spec_ctrl(tifn));
|
|
|
|
}
|
|
|
|
|
2018-11-25 19:33:35 +01:00
|
|
|
/*
|
|
|
|
* Update the MSRs managing speculation control, during context switch.
|
|
|
|
*
|
|
|
|
* tifp: Previous task's thread flags
|
|
|
|
* tifn: Next task's thread flags
|
|
|
|
*/
|
|
|
|
static __always_inline void __speculation_ctrl_update(unsigned long tifp,
|
|
|
|
unsigned long tifn)
|
2018-05-09 21:53:09 +02:00
|
|
|
{
|
2018-11-25 19:33:46 +01:00
|
|
|
unsigned long tif_diff = tifp ^ tifn;
|
2018-11-25 19:33:35 +01:00
|
|
|
u64 msr = x86_spec_ctrl_base;
|
|
|
|
bool updmsr = false;
|
|
|
|
|
2018-11-25 19:33:46 +01:00
|
|
|
/*
|
|
|
|
* If TIF_SSBD is different, select the proper mitigation
|
|
|
|
* method. Note that if SSBD mitigation is disabled or permanentely
|
|
|
|
* enabled this branch can't be taken because nothing can set
|
|
|
|
* TIF_SSBD.
|
|
|
|
*/
|
|
|
|
if (tif_diff & _TIF_SSBD) {
|
2018-11-25 19:33:35 +01:00
|
|
|
if (static_cpu_has(X86_FEATURE_VIRT_SSBD)) {
|
|
|
|
amd_set_ssb_virt_state(tifn);
|
|
|
|
} else if (static_cpu_has(X86_FEATURE_LS_CFG_SSBD)) {
|
|
|
|
amd_set_core_ssb_state(tifn);
|
|
|
|
} else if (static_cpu_has(X86_FEATURE_SPEC_CTRL_SSBD) ||
|
|
|
|
static_cpu_has(X86_FEATURE_AMD_SSBD)) {
|
|
|
|
msr |= ssbd_tif_to_spec_ctrl(tifn);
|
|
|
|
updmsr = true;
|
|
|
|
}
|
|
|
|
}
|
2018-05-09 21:53:09 +02:00
|
|
|
|
2018-11-25 19:33:46 +01:00
|
|
|
/*
|
|
|
|
* Only evaluate TIF_SPEC_IB if conditional STIBP is enabled,
|
|
|
|
* otherwise avoid the MSR write.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_SMP) &&
|
|
|
|
static_branch_unlikely(&switch_to_cond_stibp)) {
|
|
|
|
updmsr |= !!(tif_diff & _TIF_SPEC_IB);
|
|
|
|
msr |= stibp_tif_to_spec_ctrl(tifn);
|
|
|
|
}
|
|
|
|
|
2018-11-25 19:33:35 +01:00
|
|
|
if (updmsr)
|
|
|
|
wrmsrl(MSR_IA32_SPEC_CTRL, msr);
|
2018-05-09 21:53:09 +02:00
|
|
|
}
|
|
|
|
|
2018-11-28 10:56:57 +01:00
|
|
|
static unsigned long speculation_ctrl_update_tif(struct task_struct *tsk)
|
2018-05-09 21:53:09 +02:00
|
|
|
{
|
2018-11-28 10:56:57 +01:00
|
|
|
if (test_and_clear_tsk_thread_flag(tsk, TIF_SPEC_FORCE_UPDATE)) {
|
|
|
|
if (task_spec_ssb_disable(tsk))
|
|
|
|
set_tsk_thread_flag(tsk, TIF_SSBD);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(tsk, TIF_SSBD);
|
x86/speculation: Add prctl() control for indirect branch speculation
commit 9137bb27e60e554dab694eafa4cca241fa3a694f upstream.
Add the PR_SPEC_INDIRECT_BRANCH option for the PR_GET_SPECULATION_CTRL and
PR_SET_SPECULATION_CTRL prctls to allow fine grained per task control of
indirect branch speculation via STIBP and IBPB.
Invocations:
Check indirect branch speculation status with
- prctl(PR_GET_SPECULATION_CTRL, PR_SPEC_INDIRECT_BRANCH, 0, 0, 0);
Enable indirect branch speculation with
- prctl(PR_SET_SPECULATION_CTRL, PR_SPEC_INDIRECT_BRANCH, PR_SPEC_ENABLE, 0, 0);
Disable indirect branch speculation with
- prctl(PR_SET_SPECULATION_CTRL, PR_SPEC_INDIRECT_BRANCH, PR_SPEC_DISABLE, 0, 0);
Force disable indirect branch speculation with
- prctl(PR_SET_SPECULATION_CTRL, PR_SPEC_INDIRECT_BRANCH, PR_SPEC_FORCE_DISABLE, 0, 0);
See Documentation/userspace-api/spec_ctrl.rst.
Signed-off-by: Tim Chen <tim.c.chen@linux.intel.com>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Reviewed-by: Ingo Molnar <mingo@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Jiri Kosina <jkosina@suse.cz>
Cc: Tom Lendacky <thomas.lendacky@amd.com>
Cc: Josh Poimboeuf <jpoimboe@redhat.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: David Woodhouse <dwmw@amazon.co.uk>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Dave Hansen <dave.hansen@intel.com>
Cc: Casey Schaufler <casey.schaufler@intel.com>
Cc: Asit Mallick <asit.k.mallick@intel.com>
Cc: Arjan van de Ven <arjan@linux.intel.com>
Cc: Jon Masters <jcm@redhat.com>
Cc: Waiman Long <longman9394@gmail.com>
Cc: Greg KH <gregkh@linuxfoundation.org>
Cc: Dave Stewart <david.c.stewart@intel.com>
Cc: Kees Cook <keescook@chromium.org>
Link: https://lkml.kernel.org/r/20181125185005.866780996@linutronix.de
[bwh: Backported to 4.9:
- Renumber the PFA flags
- Drop changes in tools/include/uapi/linux/prctl.h
- Adjust filename]
Signed-off-by: Ben Hutchings <ben@decadent.org.uk>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2018-11-25 19:33:53 +01:00
|
|
|
|
|
|
|
if (task_spec_ib_disable(tsk))
|
|
|
|
set_tsk_thread_flag(tsk, TIF_SPEC_IB);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(tsk, TIF_SPEC_IB);
|
2018-11-28 10:56:57 +01:00
|
|
|
}
|
|
|
|
/* Return the updated threadinfo flags*/
|
|
|
|
return task_thread_info(tsk)->flags;
|
2018-05-09 21:53:09 +02:00
|
|
|
}
|
2018-04-29 15:21:42 +02:00
|
|
|
|
2018-11-25 19:33:34 +01:00
|
|
|
void speculation_ctrl_update(unsigned long tif)
|
2018-04-29 15:21:42 +02:00
|
|
|
{
|
2018-11-25 19:33:35 +01:00
|
|
|
/* Forced update. Make sure all relevant TIF flags are different */
|
2018-05-09 21:53:09 +02:00
|
|
|
preempt_disable();
|
2018-11-25 19:33:35 +01:00
|
|
|
__speculation_ctrl_update(~tif, tif);
|
2018-05-09 21:53:09 +02:00
|
|
|
preempt_enable();
|
2018-04-29 15:21:42 +02:00
|
|
|
}
|
|
|
|
|
2018-11-28 10:56:57 +01:00
|
|
|
/* Called from seccomp/prctl update */
|
|
|
|
void speculation_ctrl_update_current(void)
|
|
|
|
{
|
|
|
|
preempt_disable();
|
|
|
|
speculation_ctrl_update(speculation_ctrl_update_tif(current));
|
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
2018-11-25 19:33:47 +01:00
|
|
|
void __switch_to_xtra(struct task_struct *prev_p, struct task_struct *next_p)
|
2009-02-27 13:25:28 -08:00
|
|
|
{
|
|
|
|
struct thread_struct *prev, *next;
|
2017-02-14 00:11:02 -08:00
|
|
|
unsigned long tifp, tifn;
|
2009-02-27 13:25:28 -08:00
|
|
|
|
|
|
|
prev = &prev_p->thread;
|
|
|
|
next = &next_p->thread;
|
|
|
|
|
2017-02-14 00:11:02 -08:00
|
|
|
tifn = READ_ONCE(task_thread_info(next_p)->flags);
|
|
|
|
tifp = READ_ONCE(task_thread_info(prev_p)->flags);
|
2018-11-25 19:33:47 +01:00
|
|
|
switch_to_bitmap(prev, next, tifp, tifn);
|
2017-02-14 00:11:02 -08:00
|
|
|
|
2009-09-19 09:40:22 +03:00
|
|
|
propagate_user_return_notify(prev_p, next_p);
|
2017-02-14 00:11:02 -08:00
|
|
|
|
2017-02-14 00:11:03 -08:00
|
|
|
if ((tifp & _TIF_BLOCKSTEP || tifn & _TIF_BLOCKSTEP) &&
|
|
|
|
arch_has_block_step()) {
|
|
|
|
unsigned long debugctl, msk;
|
2010-03-25 14:51:51 +01:00
|
|
|
|
2017-02-14 00:11:03 -08:00
|
|
|
rdmsrl(MSR_IA32_DEBUGCTLMSR, debugctl);
|
2010-03-25 14:51:51 +01:00
|
|
|
debugctl &= ~DEBUGCTLMSR_BTF;
|
2017-02-14 00:11:03 -08:00
|
|
|
msk = tifn & _TIF_BLOCKSTEP;
|
|
|
|
debugctl |= (msk >> TIF_BLOCKSTEP) << DEBUGCTLMSR_BTF_SHIFT;
|
|
|
|
wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl);
|
2010-03-25 14:51:51 +01:00
|
|
|
}
|
2009-02-27 13:25:28 -08:00
|
|
|
|
2017-02-14 00:11:04 -08:00
|
|
|
if ((tifp ^ tifn) & _TIF_NOTSC)
|
|
|
|
cr4_toggle_bits(X86_CR4_TSD);
|
2018-04-29 15:21:42 +02:00
|
|
|
|
2018-11-28 10:56:57 +01:00
|
|
|
if (likely(!((tifp | tifn) & _TIF_SPEC_FORCE_UPDATE))) {
|
|
|
|
__speculation_ctrl_update(tifp, tifn);
|
|
|
|
} else {
|
|
|
|
speculation_ctrl_update_tif(prev_p);
|
|
|
|
tifn = speculation_ctrl_update_tif(next_p);
|
|
|
|
|
|
|
|
/* Enforce MSR update to ensure consistent state */
|
|
|
|
__speculation_ctrl_update(~tifn, tifn);
|
|
|
|
}
|
2009-02-27 13:25:28 -08:00
|
|
|
}
|
|
|
|
|
2008-06-09 18:35:28 +02:00
|
|
|
/*
|
|
|
|
* Idle related variables and functions
|
|
|
|
*/
|
2010-11-03 17:06:14 +01:00
|
|
|
unsigned long boot_option_idle_override = IDLE_NO_OVERRIDE;
|
2008-06-09 18:35:28 +02:00
|
|
|
EXPORT_SYMBOL(boot_option_idle_override);
|
|
|
|
|
2013-02-09 21:45:03 -05:00
|
|
|
static void (*x86_idle)(void);
|
2008-06-09 18:35:28 +02:00
|
|
|
|
2012-03-25 23:00:04 +02:00
|
|
|
#ifndef CONFIG_SMP
|
|
|
|
static inline void play_dead(void)
|
|
|
|
{
|
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
void enter_idle(void)
|
|
|
|
{
|
2012-05-11 15:35:27 +08:00
|
|
|
this_cpu_write(is_idle, 1);
|
2011-06-15 17:21:57 -07:00
|
|
|
idle_notifier_call_chain(IDLE_START);
|
2012-03-25 23:00:04 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
static void __exit_idle(void)
|
|
|
|
{
|
|
|
|
if (x86_test_and_clear_bit_percpu(0, is_idle) == 0)
|
|
|
|
return;
|
2011-06-15 17:21:57 -07:00
|
|
|
idle_notifier_call_chain(IDLE_END);
|
2012-03-25 23:00:04 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Called from interrupts to signify idle end */
|
|
|
|
void exit_idle(void)
|
|
|
|
{
|
|
|
|
/* idle loop has pid 0 */
|
|
|
|
if (current->pid)
|
|
|
|
return;
|
|
|
|
__exit_idle();
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2013-03-21 22:50:03 +01:00
|
|
|
void arch_cpu_idle_enter(void)
|
|
|
|
{
|
|
|
|
local_touch_nmi();
|
|
|
|
enter_idle();
|
|
|
|
}
|
2012-03-25 23:00:04 +02:00
|
|
|
|
2013-03-21 22:50:03 +01:00
|
|
|
void arch_cpu_idle_exit(void)
|
|
|
|
{
|
|
|
|
__exit_idle();
|
|
|
|
}
|
2012-03-25 23:00:04 +02:00
|
|
|
|
2013-03-21 22:50:03 +01:00
|
|
|
void arch_cpu_idle_dead(void)
|
|
|
|
{
|
|
|
|
play_dead();
|
|
|
|
}
|
2012-03-25 23:00:04 +02:00
|
|
|
|
2013-03-21 22:50:03 +01:00
|
|
|
/*
|
|
|
|
* Called from the generic idle code.
|
|
|
|
*/
|
|
|
|
void arch_cpu_idle(void)
|
|
|
|
{
|
2014-01-29 12:45:12 -05:00
|
|
|
x86_idle();
|
2012-03-25 23:00:04 +02:00
|
|
|
}
|
|
|
|
|
2008-06-09 18:35:28 +02:00
|
|
|
/*
|
2013-03-21 22:50:03 +01:00
|
|
|
* We use this if we don't have any better idle routine..
|
2008-06-09 18:35:28 +02:00
|
|
|
*/
|
2016-10-07 17:02:55 -07:00
|
|
|
void __cpuidle default_idle(void)
|
2008-06-09 18:35:28 +02:00
|
|
|
{
|
2012-10-25 18:13:11 +02:00
|
|
|
trace_cpu_idle_rcuidle(1, smp_processor_id());
|
2013-03-21 22:50:03 +01:00
|
|
|
safe_halt();
|
2012-10-25 18:13:11 +02:00
|
|
|
trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id());
|
2008-06-09 18:35:28 +02:00
|
|
|
}
|
2011-06-14 12:45:10 -07:00
|
|
|
#ifdef CONFIG_APM_MODULE
|
2008-06-09 18:35:28 +02:00
|
|
|
EXPORT_SYMBOL(default_idle);
|
|
|
|
#endif
|
|
|
|
|
2013-02-09 23:08:07 -05:00
|
|
|
#ifdef CONFIG_XEN
|
|
|
|
bool xen_set_default_idle(void)
|
2011-11-21 18:02:02 -05:00
|
|
|
{
|
2013-02-09 21:45:03 -05:00
|
|
|
bool ret = !!x86_idle;
|
2011-11-21 18:02:02 -05:00
|
|
|
|
2013-02-09 21:45:03 -05:00
|
|
|
x86_idle = default_idle;
|
2011-11-21 18:02:02 -05:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2013-02-09 23:08:07 -05:00
|
|
|
#endif
|
2008-11-11 14:33:44 +01:00
|
|
|
void stop_this_cpu(void *dummy)
|
|
|
|
{
|
|
|
|
local_irq_disable();
|
|
|
|
/*
|
|
|
|
* Remove this CPU:
|
|
|
|
*/
|
2009-03-13 14:49:54 +10:30
|
|
|
set_cpu_online(smp_processor_id(), false);
|
2008-11-11 14:33:44 +01:00
|
|
|
disable_local_APIC();
|
2015-08-12 18:29:40 +02:00
|
|
|
mcheck_cpu_clear(this_cpu_ptr(&cpu_info));
|
2008-11-11 14:33:44 +01:00
|
|
|
|
x86 idle: remove 32-bit-only "no-hlt" parameter, hlt_works_ok flag
Remove 32-bit x86 a cmdline param "no-hlt",
and the cpuinfo_x86.hlt_works_ok that it sets.
If a user wants to avoid HLT, then "idle=poll"
is much more useful, as it avoids invocation of HLT
in idle, while "no-hlt" failed to do so.
Indeed, hlt_works_ok was consulted in only 3 places.
First, in /proc/cpuinfo where "hlt_bug yes"
would be printed if and only if the user booted
the system with "no-hlt" -- as there was no other code
to set that flag.
Second, check_hlt() would not invoke halt() if "no-hlt"
were on the cmdline.
Third, it was consulted in stop_this_cpu(), which is invoked
by native_machine_halt()/reboot_interrupt()/smp_stop_nmi_callback() --
all cases where the machine is being shutdown/reset.
The flag was not consulted in the more frequently invoked
play_dead()/hlt_play_dead() used in processor offline and suspend.
Since Linux-3.0 there has been a run-time notice upon "no-hlt" invocations
indicating that it would be removed in 2012.
Signed-off-by: Len Brown <len.brown@intel.com>
Cc: x86@kernel.org
2013-02-10 02:28:46 -05:00
|
|
|
for (;;)
|
|
|
|
halt();
|
2008-04-25 17:39:01 +02:00
|
|
|
}
|
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
bool amd_e400_c1e_detected;
|
|
|
|
EXPORT_SYMBOL(amd_e400_c1e_detected);
|
2008-06-09 19:15:00 +02:00
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
static cpumask_var_t amd_e400_c1e_mask;
|
2008-09-22 18:54:29 +02:00
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
void amd_e400_remove_cpu(int cpu)
|
2008-09-22 18:54:29 +02:00
|
|
|
{
|
2011-04-01 16:59:53 -04:00
|
|
|
if (amd_e400_c1e_mask != NULL)
|
|
|
|
cpumask_clear_cpu(cpu, amd_e400_c1e_mask);
|
2008-09-22 18:54:29 +02:00
|
|
|
}
|
|
|
|
|
2008-06-09 19:15:00 +02:00
|
|
|
/*
|
2011-04-01 16:59:53 -04:00
|
|
|
* AMD Erratum 400 aware idle routine. We check for C1E active in the interrupt
|
2008-06-09 19:15:00 +02:00
|
|
|
* pending message MSR. If we detect C1E, then we handle it the same
|
|
|
|
* way as C3 power states (local apic timer and TSC stop)
|
|
|
|
*/
|
2011-04-01 16:59:53 -04:00
|
|
|
static void amd_e400_idle(void)
|
2008-06-09 19:15:00 +02:00
|
|
|
{
|
2011-04-01 16:59:53 -04:00
|
|
|
if (!amd_e400_c1e_detected) {
|
2008-06-09 19:15:00 +02:00
|
|
|
u32 lo, hi;
|
|
|
|
|
|
|
|
rdmsr(MSR_K8_INT_PENDING_MSG, lo, hi);
|
2010-07-27 18:53:35 +02:00
|
|
|
|
2008-06-09 19:15:00 +02:00
|
|
|
if (lo & K8_INTP_C1E_ACTIVE_MASK) {
|
2011-04-01 16:59:53 -04:00
|
|
|
amd_e400_c1e_detected = true;
|
2008-11-17 16:11:37 -08:00
|
|
|
if (!boot_cpu_has(X86_FEATURE_NONSTOP_TSC))
|
2008-09-18 21:12:10 +02:00
|
|
|
mark_tsc_unstable("TSC halt in AMD C1E");
|
2012-05-21 19:50:07 -07:00
|
|
|
pr_info("System has AMD C1E enabled\n");
|
2008-06-09 19:15:00 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
if (amd_e400_c1e_detected) {
|
2008-06-09 19:15:00 +02:00
|
|
|
int cpu = smp_processor_id();
|
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
if (!cpumask_test_cpu(cpu, amd_e400_c1e_mask)) {
|
|
|
|
cpumask_set_cpu(cpu, amd_e400_c1e_mask);
|
2015-04-03 02:01:28 +02:00
|
|
|
/* Force broadcast so ACPI can not interfere. */
|
|
|
|
tick_broadcast_force();
|
2012-05-21 19:50:07 -07:00
|
|
|
pr_info("Switch to broadcast mode on CPU%d\n", cpu);
|
2008-06-09 19:15:00 +02:00
|
|
|
}
|
2015-04-03 02:05:53 +02:00
|
|
|
tick_broadcast_enter();
|
2008-06-17 09:12:03 +02:00
|
|
|
|
2008-06-09 19:15:00 +02:00
|
|
|
default_idle();
|
2008-06-17 09:12:03 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The switch back from broadcast mode needs to be
|
|
|
|
* called with interrupts disabled.
|
|
|
|
*/
|
2013-09-11 12:43:13 +02:00
|
|
|
local_irq_disable();
|
2015-04-03 02:05:53 +02:00
|
|
|
tick_broadcast_exit();
|
2013-09-11 12:43:13 +02:00
|
|
|
local_irq_enable();
|
2008-06-09 19:15:00 +02:00
|
|
|
} else
|
|
|
|
default_idle();
|
|
|
|
}
|
|
|
|
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
/*
|
|
|
|
* Intel Core2 and older machines prefer MWAIT over HALT for C1.
|
|
|
|
* We can't rely on cpuidle installing MWAIT, because it will not load
|
|
|
|
* on systems that support only C1 -- so the boot default must be MWAIT.
|
|
|
|
*
|
|
|
|
* Some AMD machines are the opposite, they depend on using HALT.
|
|
|
|
*
|
|
|
|
* So for default C1, which is used during boot until cpuidle loads,
|
|
|
|
* use MWAIT-C1 on Intel HW that has it, else use HALT.
|
|
|
|
*/
|
|
|
|
static int prefer_mwait_c1_over_halt(const struct cpuinfo_x86 *c)
|
|
|
|
{
|
|
|
|
if (c->x86_vendor != X86_VENDOR_INTEL)
|
|
|
|
return 0;
|
|
|
|
|
2016-07-18 11:41:10 -07:00
|
|
|
if (!cpu_has(c, X86_FEATURE_MWAIT) || static_cpu_has_bug(X86_BUG_MONITOR))
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2015-05-26 10:28:09 +02:00
|
|
|
* MONITOR/MWAIT with no hints, used for default C1 state. This invokes MWAIT
|
|
|
|
* with interrupts enabled and no flags, which is backwards compatible with the
|
|
|
|
* original MWAIT implementation.
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
*/
|
2016-10-07 17:02:55 -07:00
|
|
|
static __cpuidle void mwait_idle(void)
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
{
|
2014-01-18 17:14:44 +01:00
|
|
|
if (!current_set_polling_and_test()) {
|
2015-08-20 12:54:39 +08:00
|
|
|
trace_cpu_idle_rcuidle(1, smp_processor_id());
|
2014-01-18 17:14:44 +01:00
|
|
|
if (this_cpu_has(X86_BUG_CLFLUSH_MONITOR)) {
|
2016-01-28 19:02:51 +02:00
|
|
|
mb(); /* quirk */
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
clflush((void *)¤t_thread_info()->flags);
|
2016-01-28 19:02:51 +02:00
|
|
|
mb(); /* quirk */
|
2014-01-18 17:14:44 +01:00
|
|
|
}
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
|
|
|
|
__monitor((void *)¤t_thread_info()->flags, 0, 0);
|
|
|
|
if (!need_resched())
|
|
|
|
__sti_mwait(0, 0);
|
|
|
|
else
|
|
|
|
local_irq_enable();
|
2015-08-20 12:54:39 +08:00
|
|
|
trace_cpu_idle_rcuidle(PWR_EVENT_EXIT, smp_processor_id());
|
2014-01-18 17:14:44 +01:00
|
|
|
} else {
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
local_irq_enable();
|
2014-01-18 17:14:44 +01:00
|
|
|
}
|
|
|
|
__current_clr_polling();
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
}
|
|
|
|
|
x86: delete __cpuinit usage from all x86 files
The __cpuinit type of throwaway sections might have made sense
some time ago when RAM was more constrained, but now the savings
do not offset the cost and complications. For example, the fix in
commit 5e427ec2d0 ("x86: Fix bit corruption at CPU resume time")
is a good example of the nasty type of bugs that can be created
with improper use of the various __init prefixes.
After a discussion on LKML[1] it was decided that cpuinit should go
the way of devinit and be phased out. Once all the users are gone,
we can then finally remove the macros themselves from linux/init.h.
Note that some harmless section mismatch warnings may result, since
notify_cpu_starting() and cpu_up() are arch independent (kernel/cpu.c)
are flagged as __cpuinit -- so if we remove the __cpuinit from
arch specific callers, we will also get section mismatch warnings.
As an intermediate step, we intend to turn the linux/init.h cpuinit
content into no-ops as early as possible, since that will get rid
of these warnings. In any case, they are temporary and harmless.
This removes all the arch/x86 uses of the __cpuinit macros from
all C files. x86 only had the one __CPUINIT used in assembly files,
and it wasn't paired off with a .previous or a __FINIT, so we can
delete it directly w/o any corresponding additional change there.
[1] https://lkml.org/lkml/2013/5/20/589
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: x86@kernel.org
Acked-by: Ingo Molnar <mingo@kernel.org>
Acked-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: H. Peter Anvin <hpa@linux.intel.com>
Signed-off-by: Paul Gortmaker <paul.gortmaker@windriver.com>
2013-06-18 18:23:59 -04:00
|
|
|
void select_idle_routine(const struct cpuinfo_x86 *c)
|
2008-04-25 17:39:01 +02:00
|
|
|
{
|
2009-01-27 17:07:08 +01:00
|
|
|
#ifdef CONFIG_SMP
|
2013-03-21 22:50:03 +01:00
|
|
|
if (boot_option_idle_override == IDLE_POLL && smp_num_siblings > 1)
|
2012-05-21 19:50:07 -07:00
|
|
|
pr_warn_once("WARNING: polling idle and HT enabled, performance may degrade\n");
|
2008-04-25 17:39:01 +02:00
|
|
|
#endif
|
2013-03-21 22:50:03 +01:00
|
|
|
if (x86_idle || boot_option_idle_override == IDLE_POLL)
|
2008-06-09 16:59:53 +02:00
|
|
|
return;
|
|
|
|
|
2016-12-09 19:29:09 +01:00
|
|
|
if (boot_cpu_has_bug(X86_BUG_AMD_E400)) {
|
2012-05-21 19:50:07 -07:00
|
|
|
pr_info("using AMD E400 aware idle routine\n");
|
2013-02-09 21:45:03 -05:00
|
|
|
x86_idle = amd_e400_idle;
|
sched/idle/x86: Restore mwait_idle() to fix boot hangs, to improve power savings and to improve performance
In Linux-3.9 we removed the mwait_idle() loop:
69fb3676df33 ("x86 idle: remove mwait_idle() and "idle=mwait" cmdline param")
The reasoning was that modern machines should be sufficiently
happy during the boot process using the default_idle() HALT
loop, until cpuidle loads and either acpi_idle or intel_idle
invoke the newer MWAIT-with-hints idle loop.
But two machines reported problems:
1. Certain Core2-era machines support MWAIT-C1 and HALT only.
MWAIT-C1 is preferred for optimal power and performance.
But if they support just C1, cpuidle never loads and
so they use the boot-time default idle loop forever.
2. Some laptops will boot-hang if HALT is used,
but will boot successfully if MWAIT is used.
This appears to be a hidden assumption in BIOS SMI,
that is presumably valid on the proprietary OS
where the BIOS was validated.
https://bugzilla.kernel.org/show_bug.cgi?id=60770
So here we effectively revert the patch above, restoring
the mwait_idle() loop. However, we don't bother restoring
the idle=mwait cmdline parameter, since it appears to add
no value.
Maintainer notes:
For 3.9, simply revert 69fb3676df
for 3.10, patch -F3 applies, fuzz needed due to __cpuinit use in
context For 3.11, 3.12, 3.13, this patch applies cleanly
Tested-by: Mike Galbraith <bitbucket@online.de>
Signed-off-by: Len Brown <len.brown@intel.com>
Acked-by: Mike Galbraith <bitbucket@online.de>
Cc: <stable@vger.kernel.org> # 3.9+
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Ian Malone <ibmalone@gmail.com>
Cc: Josh Boyer <jwboyer@redhat.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/345254a551eb5a6a866e048d7ab570fd2193aca4.1389763084.git.len.brown@intel.com
[ Ported to recent kernels. ]
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-01-15 00:37:34 -05:00
|
|
|
} else if (prefer_mwait_c1_over_halt(c)) {
|
|
|
|
pr_info("using mwait in idle threads\n");
|
|
|
|
x86_idle = mwait_idle;
|
2008-06-09 16:59:53 +02:00
|
|
|
} else
|
2013-02-09 21:45:03 -05:00
|
|
|
x86_idle = default_idle;
|
2008-04-25 17:39:01 +02:00
|
|
|
}
|
|
|
|
|
2011-04-01 16:59:53 -04:00
|
|
|
void __init init_amd_e400_c1e_mask(void)
|
2009-03-17 14:50:34 +10:30
|
|
|
{
|
2011-04-01 16:59:53 -04:00
|
|
|
/* If we're using amd_e400_idle, we need to allocate amd_e400_c1e_mask. */
|
2013-02-09 21:45:03 -05:00
|
|
|
if (x86_idle == amd_e400_idle)
|
2011-04-01 16:59:53 -04:00
|
|
|
zalloc_cpumask_var(&amd_e400_c1e_mask, GFP_KERNEL);
|
2009-03-17 14:50:34 +10:30
|
|
|
}
|
|
|
|
|
2008-04-25 17:39:01 +02:00
|
|
|
static int __init idle_setup(char *str)
|
|
|
|
{
|
2008-07-05 15:53:36 +04:00
|
|
|
if (!str)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-04-25 17:39:01 +02:00
|
|
|
if (!strcmp(str, "poll")) {
|
2012-05-21 19:50:07 -07:00
|
|
|
pr_info("using polling idle threads\n");
|
2010-11-03 17:06:14 +01:00
|
|
|
boot_option_idle_override = IDLE_POLL;
|
2013-03-21 22:50:03 +01:00
|
|
|
cpu_idle_poll_ctrl(true);
|
2010-11-03 17:06:14 +01:00
|
|
|
} else if (!strcmp(str, "halt")) {
|
2008-06-24 17:58:53 +08:00
|
|
|
/*
|
|
|
|
* When the boot option of idle=halt is added, halt is
|
|
|
|
* forced to be used for CPU idle. In such case CPU C2/C3
|
|
|
|
* won't be used again.
|
|
|
|
* To continue to load the CPU idle driver, don't touch
|
|
|
|
* the boot_option_idle_override.
|
|
|
|
*/
|
2013-02-09 21:45:03 -05:00
|
|
|
x86_idle = default_idle;
|
2010-11-03 17:06:14 +01:00
|
|
|
boot_option_idle_override = IDLE_HALT;
|
2008-06-24 18:01:09 +08:00
|
|
|
} else if (!strcmp(str, "nomwait")) {
|
|
|
|
/*
|
|
|
|
* If the boot option of "idle=nomwait" is added,
|
|
|
|
* it means that mwait will be disabled for CPU C2/C3
|
|
|
|
* states. In such case it won't touch the variable
|
|
|
|
* of boot_option_idle_override.
|
|
|
|
*/
|
2010-11-03 17:06:14 +01:00
|
|
|
boot_option_idle_override = IDLE_NOMWAIT;
|
2008-06-24 17:58:53 +08:00
|
|
|
} else
|
2008-04-25 17:39:01 +02:00
|
|
|
return -1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("idle", idle_setup);
|
|
|
|
|
2009-05-11 22:05:28 -04:00
|
|
|
unsigned long arch_align_stack(unsigned long sp)
|
|
|
|
{
|
|
|
|
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
|
|
|
|
sp -= get_random_int() % 8192;
|
|
|
|
return sp & ~0xf;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned long arch_randomize_brk(struct mm_struct *mm)
|
|
|
|
{
|
2016-10-11 13:53:56 -07:00
|
|
|
return randomize_page(mm->brk, 0x02000000);
|
2009-05-11 22:05:28 -04:00
|
|
|
}
|
|
|
|
|
2016-08-13 12:38:21 -04:00
|
|
|
/*
|
|
|
|
* Return saved PC of a blocked thread.
|
|
|
|
* What is this good for? it will be always the scheduler or ret_from_fork.
|
|
|
|
*/
|
|
|
|
unsigned long thread_saved_pc(struct task_struct *tsk)
|
|
|
|
{
|
|
|
|
struct inactive_task_frame *frame =
|
|
|
|
(struct inactive_task_frame *) READ_ONCE(tsk->thread.sp);
|
|
|
|
return READ_ONCE_NOCHECK(frame->ret_addr);
|
|
|
|
}
|
|
|
|
|
2015-09-30 08:38:23 +00:00
|
|
|
/*
|
|
|
|
* Called from fs/proc with a reference on @p to find the function
|
|
|
|
* which called into schedule(). This needs to be done carefully
|
|
|
|
* because the task might wake up and we might look at a stack
|
|
|
|
* changing under us.
|
|
|
|
*/
|
|
|
|
unsigned long get_wchan(struct task_struct *p)
|
|
|
|
{
|
2016-09-15 22:45:46 -07:00
|
|
|
unsigned long start, bottom, top, sp, fp, ip, ret = 0;
|
2015-09-30 08:38:23 +00:00
|
|
|
int count = 0;
|
|
|
|
|
|
|
|
if (!p || p == current || p->state == TASK_RUNNING)
|
|
|
|
return 0;
|
|
|
|
|
2016-09-15 22:45:46 -07:00
|
|
|
if (!try_get_task_stack(p))
|
|
|
|
return 0;
|
|
|
|
|
2015-09-30 08:38:23 +00:00
|
|
|
start = (unsigned long)task_stack_page(p);
|
|
|
|
if (!start)
|
2016-09-15 22:45:46 -07:00
|
|
|
goto out;
|
2015-09-30 08:38:23 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Layout of the stack page:
|
|
|
|
*
|
|
|
|
* ----------- topmax = start + THREAD_SIZE - sizeof(unsigned long)
|
|
|
|
* PADDING
|
|
|
|
* ----------- top = topmax - TOP_OF_KERNEL_STACK_PADDING
|
|
|
|
* stack
|
2016-09-13 14:29:25 -07:00
|
|
|
* ----------- bottom = start
|
2015-09-30 08:38:23 +00:00
|
|
|
*
|
|
|
|
* The tasks stack pointer points at the location where the
|
|
|
|
* framepointer is stored. The data on the stack is:
|
|
|
|
* ... IP FP ... IP FP
|
|
|
|
*
|
|
|
|
* We need to read FP and IP, so we need to adjust the upper
|
|
|
|
* bound by another unsigned long.
|
|
|
|
*/
|
|
|
|
top = start + THREAD_SIZE - TOP_OF_KERNEL_STACK_PADDING;
|
|
|
|
top -= 2 * sizeof(unsigned long);
|
2016-09-13 14:29:25 -07:00
|
|
|
bottom = start;
|
2015-09-30 08:38:23 +00:00
|
|
|
|
|
|
|
sp = READ_ONCE(p->thread.sp);
|
|
|
|
if (sp < bottom || sp > top)
|
2016-09-15 22:45:46 -07:00
|
|
|
goto out;
|
2015-09-30 08:38:23 +00:00
|
|
|
|
2016-08-13 12:38:18 -04:00
|
|
|
fp = READ_ONCE_NOCHECK(((struct inactive_task_frame *)sp)->bp);
|
2015-09-30 08:38:23 +00:00
|
|
|
do {
|
|
|
|
if (fp < bottom || fp > top)
|
2016-09-15 22:45:46 -07:00
|
|
|
goto out;
|
2015-10-19 11:37:18 +03:00
|
|
|
ip = READ_ONCE_NOCHECK(*(unsigned long *)(fp + sizeof(unsigned long)));
|
2016-09-15 22:45:46 -07:00
|
|
|
if (!in_sched_functions(ip)) {
|
|
|
|
ret = ip;
|
|
|
|
goto out;
|
|
|
|
}
|
2015-10-19 11:37:18 +03:00
|
|
|
fp = READ_ONCE_NOCHECK(*(unsigned long *)fp);
|
2015-09-30 08:38:23 +00:00
|
|
|
} while (count++ < 16 && p->state != TASK_RUNNING);
|
2016-09-15 22:45:46 -07:00
|
|
|
|
|
|
|
out:
|
|
|
|
put_task_stack(p);
|
|
|
|
return ret;
|
2015-09-30 08:38:23 +00:00
|
|
|
}
|