e9623b3559
The previous revert of0c07ee38c9
left out the mwait disable condition for AMD family 10H/11H CPUs. Andreas Herrman said: It depends on the CPU. For AMD CPUs that support MWAIT this is wrong. Family 0x10 and 0x11 CPUs will enter C1 on HLT. Powersavings then depend on a clock divisor and current Pstate of the core. If all cores of a processor are in halt state (C1) the processor can enter the C1E (C1 enhanced) state. If mwait is used this will never happen. Thus HLT saves more power than MWAIT here. It might be best to switch off the mwait flag for these AMD CPU families like it was introduced with commitf039b75471
(x86: Don't use MWAIT on AMD Family 10) Re-add the AMD families 10H/11H check and disable the mwait usage for those. Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
179 lines
4.3 KiB
C
179 lines
4.3 KiB
C
#include <linux/errno.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pm.h>
|
|
|
|
struct kmem_cache *task_xstate_cachep;
|
|
|
|
int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src)
|
|
{
|
|
*dst = *src;
|
|
if (src->thread.xstate) {
|
|
dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep,
|
|
GFP_KERNEL);
|
|
if (!dst->thread.xstate)
|
|
return -ENOMEM;
|
|
WARN_ON((unsigned long)dst->thread.xstate & 15);
|
|
memcpy(dst->thread.xstate, src->thread.xstate, xstate_size);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void free_thread_xstate(struct task_struct *tsk)
|
|
{
|
|
if (tsk->thread.xstate) {
|
|
kmem_cache_free(task_xstate_cachep, tsk->thread.xstate);
|
|
tsk->thread.xstate = NULL;
|
|
}
|
|
}
|
|
|
|
void free_thread_info(struct thread_info *ti)
|
|
{
|
|
free_thread_xstate(ti->task);
|
|
free_pages((unsigned long)ti, get_order(THREAD_SIZE));
|
|
}
|
|
|
|
void arch_task_cache_init(void)
|
|
{
|
|
task_xstate_cachep =
|
|
kmem_cache_create("task_xstate", xstate_size,
|
|
__alignof__(union thread_xstate),
|
|
SLAB_PANIC, NULL);
|
|
}
|
|
|
|
static void do_nothing(void *unused)
|
|
{
|
|
}
|
|
|
|
/*
|
|
* cpu_idle_wait - Used to ensure that all the CPUs discard old value of
|
|
* pm_idle and update to new pm_idle value. Required while changing pm_idle
|
|
* handler on SMP systems.
|
|
*
|
|
* Caller must have changed pm_idle to the new value before the call. Old
|
|
* pm_idle value will not be used by any CPU after the return of this function.
|
|
*/
|
|
void cpu_idle_wait(void)
|
|
{
|
|
smp_mb();
|
|
/* kick all the CPUs so that they exit out of pm_idle */
|
|
smp_call_function(do_nothing, NULL, 0, 1);
|
|
}
|
|
EXPORT_SYMBOL_GPL(cpu_idle_wait);
|
|
|
|
/*
|
|
* This uses new MONITOR/MWAIT instructions on P4 processors with PNI,
|
|
* which can obviate IPI to trigger checking of need_resched.
|
|
* We execute MONITOR against need_resched and enter optimized wait state
|
|
* through MWAIT. Whenever someone changes need_resched, we would be woken
|
|
* up from MWAIT (without an IPI).
|
|
*
|
|
* New with Core Duo processors, MWAIT can take some hints based on CPU
|
|
* capability.
|
|
*/
|
|
void mwait_idle_with_hints(unsigned long ax, unsigned long cx)
|
|
{
|
|
if (!need_resched()) {
|
|
__monitor((void *)¤t_thread_info()->flags, 0, 0);
|
|
smp_mb();
|
|
if (!need_resched())
|
|
__mwait(ax, cx);
|
|
}
|
|
}
|
|
|
|
/* Default MONITOR/MWAIT with no hints, used for default C1 state */
|
|
static void mwait_idle(void)
|
|
{
|
|
if (!need_resched()) {
|
|
__monitor((void *)¤t_thread_info()->flags, 0, 0);
|
|
smp_mb();
|
|
if (!need_resched())
|
|
__sti_mwait(0, 0);
|
|
else
|
|
local_irq_enable();
|
|
} else
|
|
local_irq_enable();
|
|
}
|
|
|
|
/*
|
|
* On SMP it's slightly faster (but much more power-consuming!)
|
|
* to poll the ->work.need_resched flag instead of waiting for the
|
|
* cross-CPU IPI to arrive. Use this option with caution.
|
|
*/
|
|
static void poll_idle(void)
|
|
{
|
|
local_irq_enable();
|
|
cpu_relax();
|
|
}
|
|
|
|
/*
|
|
* mwait selection logic:
|
|
*
|
|
* It depends on the CPU. For AMD CPUs that support MWAIT this is
|
|
* wrong. Family 0x10 and 0x11 CPUs will enter C1 on HLT. Powersavings
|
|
* then depend on a clock divisor and current Pstate of the core. If
|
|
* all cores of a processor are in halt state (C1) the processor can
|
|
* enter the C1E (C1 enhanced) state. If mwait is used this will never
|
|
* happen.
|
|
*
|
|
* idle=mwait overrides this decision and forces the usage of mwait.
|
|
*/
|
|
static int __cpuinit mwait_usable(const struct cpuinfo_x86 *c)
|
|
{
|
|
if (force_mwait)
|
|
return 1;
|
|
|
|
if (c->x86_vendor == X86_VENDOR_AMD) {
|
|
switch(c->x86) {
|
|
case 0x10:
|
|
case 0x11:
|
|
return 0;
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c)
|
|
{
|
|
static int selected;
|
|
|
|
if (selected)
|
|
return;
|
|
#ifdef CONFIG_X86_SMP
|
|
if (pm_idle == poll_idle && smp_num_siblings > 1) {
|
|
printk(KERN_WARNING "WARNING: polling idle and HT enabled,"
|
|
" performance may degrade.\n");
|
|
}
|
|
#endif
|
|
if (cpu_has(c, X86_FEATURE_MWAIT) && mwait_usable(c)) {
|
|
/*
|
|
* Skip, if setup has overridden idle.
|
|
* One CPU supports mwait => All CPUs supports mwait
|
|
*/
|
|
if (!pm_idle) {
|
|
printk(KERN_INFO "using mwait in idle threads.\n");
|
|
pm_idle = mwait_idle;
|
|
}
|
|
}
|
|
selected = 1;
|
|
}
|
|
|
|
static int __init idle_setup(char *str)
|
|
{
|
|
if (!strcmp(str, "poll")) {
|
|
printk("using polling idle threads.\n");
|
|
pm_idle = poll_idle;
|
|
} else if (!strcmp(str, "mwait"))
|
|
force_mwait = 1;
|
|
else
|
|
return -1;
|
|
|
|
boot_option_idle_override = 1;
|
|
return 0;
|
|
}
|
|
early_param("idle", idle_setup);
|
|
|