Merge: update cpuidle to match Linux v6.12
MR: https://gitlab.com/redhat/centos-stream/src/kernel/centos-stream-9/-/merge_requests/5570 JIRA: https://issues.redhat.com/browse/RHEL-64032 Signed-off-by: Mark Langsdorf <mlangsdo@redhat.com> Approved-by: Herton R. Krzesinski <herton@redhat.com> Approved-by: David Arcari <darcari@redhat.com> Approved-by: Tony Camuso <tcamuso@redhat.com> Approved-by: CKI KWF Bot <cki-ci-bot+kwf-gitlab-com@redhat.com> Merged-by: Patrick Talbert <ptalbert@redhat.com>
This commit is contained in:
commit
4cdb2b3206
|
@ -439,13 +439,8 @@ static int cpuidle_coupled_clear_pokes(int cpu)
|
|||
|
||||
static bool cpuidle_coupled_any_pokes_pending(struct cpuidle_coupled *coupled)
|
||||
{
|
||||
cpumask_t cpus;
|
||||
int ret;
|
||||
|
||||
cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus);
|
||||
ret = cpumask_and(&cpus, &cpuidle_coupled_poke_pending, &cpus);
|
||||
|
||||
return ret;
|
||||
return cpumask_first_and_and(cpu_online_mask, &coupled->coupled_cpus,
|
||||
&cpuidle_coupled_poke_pending) < nr_cpu_ids;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -626,9 +621,7 @@ out:
|
|||
|
||||
static void cpuidle_coupled_update_online_cpus(struct cpuidle_coupled *coupled)
|
||||
{
|
||||
cpumask_t cpus;
|
||||
cpumask_and(&cpus, cpu_online_mask, &coupled->coupled_cpus);
|
||||
coupled->online_count = cpumask_weight(&cpus);
|
||||
coupled->online_count = cpumask_weight_and(cpu_online_mask, &coupled->coupled_cpus);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -25,13 +25,12 @@ MODULE_PARM_DESC(force, "Load unconditionally");
|
|||
static struct cpuidle_device __percpu *haltpoll_cpuidle_devices;
|
||||
static enum cpuhp_state haltpoll_hp_state;
|
||||
|
||||
static int default_enter_idle(struct cpuidle_device *dev,
|
||||
struct cpuidle_driver *drv, int index)
|
||||
static __cpuidle int default_enter_idle(struct cpuidle_device *dev,
|
||||
struct cpuidle_driver *drv, int index)
|
||||
{
|
||||
if (current_clr_polling_and_test()) {
|
||||
local_irq_enable();
|
||||
if (current_clr_polling_and_test())
|
||||
return index;
|
||||
}
|
||||
|
||||
arch_cpu_idle();
|
||||
return index;
|
||||
}
|
||||
|
@ -142,5 +141,6 @@ static void __exit haltpoll_exit(void)
|
|||
|
||||
module_init(haltpoll_init);
|
||||
module_exit(haltpoll_exit);
|
||||
MODULE_DESCRIPTION("cpuidle driver for haltpoll governor");
|
||||
MODULE_LICENSE("GPL");
|
||||
MODULE_AUTHOR("Marcelo Tosatti <mtosatti@redhat.com>");
|
||||
|
|
|
@ -20,6 +20,7 @@
|
|||
#include <linux/string.h>
|
||||
|
||||
#include "cpuidle-psci.h"
|
||||
#include "dt_idle_genpd.h"
|
||||
|
||||
struct psci_pd_provider {
|
||||
struct list_head link;
|
||||
|
@ -66,12 +67,16 @@ static int psci_pd_init(struct device_node *np, bool use_osi)
|
|||
|
||||
/*
|
||||
* Allow power off when OSI has been successfully enabled.
|
||||
* PREEMPT_RT is not yet ready to enter domain idle states.
|
||||
* On a PREEMPT_RT based configuration the domain idle states are
|
||||
* supported, but only during system-wide suspend.
|
||||
*/
|
||||
if (use_osi && !IS_ENABLED(CONFIG_PREEMPT_RT))
|
||||
if (use_osi) {
|
||||
pd->power_off = psci_pd_power_off;
|
||||
else
|
||||
if (IS_ENABLED(CONFIG_PREEMPT_RT))
|
||||
pd->flags |= GENPD_FLAG_RPM_ALWAYS_ON;
|
||||
} else {
|
||||
pd->flags |= GENPD_FLAG_ALWAYS_ON;
|
||||
}
|
||||
|
||||
/* Use governor for CPU PM domains if it has some states to manage. */
|
||||
pd_gov = pd->states ? &pm_domain_cpu_gov : NULL;
|
||||
|
@ -137,7 +142,6 @@ static const struct of_device_id psci_of_match[] = {
|
|||
static int psci_cpuidle_domain_probe(struct platform_device *pdev)
|
||||
{
|
||||
struct device_node *np = pdev->dev.of_node;
|
||||
struct device_node *node;
|
||||
bool use_osi = psci_has_osi_support();
|
||||
int ret = 0, pd_count = 0;
|
||||
|
||||
|
@ -148,15 +152,13 @@ static int psci_cpuidle_domain_probe(struct platform_device *pdev)
|
|||
* Parse child nodes for the "#power-domain-cells" property and
|
||||
* initialize a genpd/genpd-of-provider pair when it's found.
|
||||
*/
|
||||
for_each_child_of_node(np, node) {
|
||||
for_each_child_of_node_scoped(np, node) {
|
||||
if (!of_property_present(node, "#power-domain-cells"))
|
||||
continue;
|
||||
|
||||
ret = psci_pd_init(node, use_osi);
|
||||
if (ret) {
|
||||
of_node_put(node);
|
||||
if (ret)
|
||||
goto exit;
|
||||
}
|
||||
|
||||
pd_count++;
|
||||
}
|
||||
|
@ -200,4 +202,4 @@ static int __init psci_idle_init_domains(void)
|
|||
{
|
||||
return platform_driver_register(&psci_cpuidle_domain_driver);
|
||||
}
|
||||
subsys_initcall(psci_idle_init_domains);
|
||||
core_initcall(psci_idle_init_domains);
|
||||
|
|
|
@ -28,6 +28,7 @@
|
|||
|
||||
#include "cpuidle-psci.h"
|
||||
#include "dt_idle_states.h"
|
||||
#include "dt_idle_genpd.h"
|
||||
|
||||
struct psci_cpuidle_data {
|
||||
u32 *psci_states;
|
||||
|
@ -36,6 +37,7 @@ struct psci_cpuidle_data {
|
|||
|
||||
static DEFINE_PER_CPU_READ_MOSTLY(struct psci_cpuidle_data, psci_cpuidle_data);
|
||||
static DEFINE_PER_CPU(u32, domain_state);
|
||||
static bool psci_cpuidle_use_syscore;
|
||||
static bool psci_cpuidle_use_cpuhp;
|
||||
|
||||
void psci_set_domain_state(u32 state)
|
||||
|
@ -165,6 +167,12 @@ static struct syscore_ops psci_idle_syscore_ops = {
|
|||
.resume = psci_idle_syscore_resume,
|
||||
};
|
||||
|
||||
static void psci_idle_init_syscore(void)
|
||||
{
|
||||
if (psci_cpuidle_use_syscore)
|
||||
register_syscore_ops(&psci_idle_syscore_ops);
|
||||
}
|
||||
|
||||
static void psci_idle_init_cpuhp(void)
|
||||
{
|
||||
int err;
|
||||
|
@ -172,8 +180,6 @@ static void psci_idle_init_cpuhp(void)
|
|||
if (!psci_cpuidle_use_cpuhp)
|
||||
return;
|
||||
|
||||
register_syscore_ops(&psci_idle_syscore_ops);
|
||||
|
||||
err = cpuhp_setup_state_nocalls(CPUHP_AP_CPU_PM_STARTING,
|
||||
"cpuidle/psci:online",
|
||||
psci_idle_cpuhp_up,
|
||||
|
@ -221,22 +227,23 @@ static int psci_dt_cpu_init_topology(struct cpuidle_driver *drv,
|
|||
if (!psci_has_osi_support())
|
||||
return 0;
|
||||
|
||||
if (IS_ENABLED(CONFIG_PREEMPT_RT))
|
||||
return 0;
|
||||
|
||||
data->dev = psci_dt_attach_cpu(cpu);
|
||||
data->dev = dt_idle_attach_cpu(cpu, "psci");
|
||||
if (IS_ERR_OR_NULL(data->dev))
|
||||
return PTR_ERR_OR_ZERO(data->dev);
|
||||
|
||||
psci_cpuidle_use_syscore = true;
|
||||
|
||||
/*
|
||||
* Using the deepest state for the CPU to trigger a potential selection
|
||||
* of a shared state for the domain, assumes the domain states are all
|
||||
* deeper states.
|
||||
* deeper states. On PREEMPT_RT the hierarchical topology is limited to
|
||||
* s2ram and s2idle.
|
||||
*/
|
||||
drv->states[state_count - 1].flags |= CPUIDLE_FLAG_RCU_IDLE;
|
||||
drv->states[state_count - 1].enter = psci_enter_domain_idle_state;
|
||||
drv->states[state_count - 1].enter_s2idle = psci_enter_s2idle_domain_idle_state;
|
||||
psci_cpuidle_use_cpuhp = true;
|
||||
if (!IS_ENABLED(CONFIG_PREEMPT_RT)) {
|
||||
drv->states[state_count - 1].enter = psci_enter_domain_idle_state;
|
||||
psci_cpuidle_use_cpuhp = true;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -311,7 +318,8 @@ static void psci_cpu_deinit_idle(int cpu)
|
|||
{
|
||||
struct psci_cpuidle_data *data = per_cpu_ptr(&psci_cpuidle_data, cpu);
|
||||
|
||||
psci_dt_detach_cpu(data->dev);
|
||||
dt_idle_detach_cpu(data->dev);
|
||||
psci_cpuidle_use_syscore = false;
|
||||
psci_cpuidle_use_cpuhp = false;
|
||||
}
|
||||
|
||||
|
@ -408,6 +416,7 @@ static int psci_cpuidle_probe(struct platform_device *pdev)
|
|||
goto out_fail;
|
||||
}
|
||||
|
||||
psci_idle_init_syscore();
|
||||
psci_idle_init_cpuhp();
|
||||
return 0;
|
||||
|
||||
|
|
|
@ -3,29 +3,9 @@
|
|||
#ifndef __CPUIDLE_PSCI_H
|
||||
#define __CPUIDLE_PSCI_H
|
||||
|
||||
struct device;
|
||||
struct device_node;
|
||||
|
||||
void psci_set_domain_state(u32 state);
|
||||
int psci_dt_parse_state_node(struct device_node *np, u32 *state);
|
||||
|
||||
#ifdef CONFIG_ARM_PSCI_CPUIDLE_DOMAIN
|
||||
|
||||
#include "dt_idle_genpd.h"
|
||||
|
||||
static inline struct device *psci_dt_attach_cpu(int cpu)
|
||||
{
|
||||
return dt_idle_attach_cpu(cpu, "psci");
|
||||
}
|
||||
|
||||
static inline void psci_dt_detach_cpu(struct device *dev)
|
||||
{
|
||||
dt_idle_detach_cpu(dev);
|
||||
}
|
||||
|
||||
#else
|
||||
static inline struct device *psci_dt_attach_cpu(int cpu) { return NULL; }
|
||||
static inline void psci_dt_detach_cpu(struct device *dev) { }
|
||||
#endif
|
||||
|
||||
#endif /* __CPUIDLE_PSCI_H */
|
||||
|
|
|
@ -407,7 +407,7 @@ static void __init fixup_cede0_latency(void)
|
|||
* pseries_idle_probe()
|
||||
* Choose state table for shared versus dedicated partition
|
||||
*/
|
||||
static int pseries_idle_probe(void)
|
||||
static int __init pseries_idle_probe(void)
|
||||
{
|
||||
|
||||
if (cpuidle_disable != IDLE_NO_OVERRIDE)
|
||||
|
|
|
@ -228,10 +228,7 @@ noinstr int cpuidle_enter_state(struct cpuidle_device *dev,
|
|||
if (broadcast && tick_broadcast_enter()) {
|
||||
index = find_deepest_state(drv, dev, target_state->exit_latency_ns,
|
||||
CPUIDLE_FLAG_TIMER_STOP, false);
|
||||
if (index < 0) {
|
||||
default_idle_call();
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
target_state = &drv->states[index];
|
||||
broadcast = false;
|
||||
}
|
||||
|
|
|
@ -16,6 +16,7 @@
|
|||
#include <linux/cpumask.h>
|
||||
#include <linux/tick.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/math64.h>
|
||||
|
||||
#include "cpuidle.h"
|
||||
|
||||
|
@ -187,7 +188,7 @@ static void __cpuidle_driver_init(struct cpuidle_driver *drv)
|
|||
s->target_residency = div_u64(s->target_residency_ns, NSEC_PER_USEC);
|
||||
|
||||
if (s->exit_latency > 0)
|
||||
s->exit_latency_ns = s->exit_latency * NSEC_PER_USEC;
|
||||
s->exit_latency_ns = mul_u32_u32(s->exit_latency, NSEC_PER_USEC);
|
||||
else if (s->exit_latency_ns < 0)
|
||||
s->exit_latency_ns = 0;
|
||||
else
|
||||
|
|
|
@ -130,11 +130,10 @@ out:
|
|||
|
||||
int dt_idle_pd_init_topology(struct device_node *np)
|
||||
{
|
||||
struct device_node *node;
|
||||
struct of_phandle_args child, parent;
|
||||
int ret;
|
||||
|
||||
for_each_child_of_node(np, node) {
|
||||
for_each_child_of_node_scoped(np, node) {
|
||||
if (of_parse_phandle_with_args(node, "power-domains",
|
||||
"#power-domain-cells", 0, &parent))
|
||||
continue;
|
||||
|
@ -143,10 +142,8 @@ int dt_idle_pd_init_topology(struct device_node *np)
|
|||
child.args_count = 0;
|
||||
ret = of_genpd_add_subdomain(&parent, &child);
|
||||
of_node_put(parent.np);
|
||||
if (ret) {
|
||||
of_node_put(node);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
@ -154,11 +151,10 @@ int dt_idle_pd_init_topology(struct device_node *np)
|
|||
|
||||
int dt_idle_pd_remove_topology(struct device_node *np)
|
||||
{
|
||||
struct device_node *node;
|
||||
struct of_phandle_args child, parent;
|
||||
int ret;
|
||||
|
||||
for_each_child_of_node(np, node) {
|
||||
for_each_child_of_node_scoped(np, node) {
|
||||
if (of_parse_phandle_with_args(node, "power-domains",
|
||||
"#power-domain-cells", 0, &parent))
|
||||
continue;
|
||||
|
@ -167,10 +163,8 @@ int dt_idle_pd_remove_topology(struct device_node *np)
|
|||
child.args_count = 0;
|
||||
ret = of_genpd_remove_subdomain(&parent, &child);
|
||||
of_node_put(parent.np);
|
||||
if (ret) {
|
||||
of_node_put(node);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
|
|
@ -98,10 +98,15 @@ static void adjust_poll_limit(struct cpuidle_device *dev, u64 block_ns)
|
|||
unsigned int shrink = guest_halt_poll_shrink;
|
||||
|
||||
val = dev->poll_limit_ns;
|
||||
if (shrink == 0)
|
||||
if (shrink == 0) {
|
||||
val = 0;
|
||||
else
|
||||
} else {
|
||||
val /= shrink;
|
||||
/* Reset value to 0 if shrunk below grow_start */
|
||||
if (val < guest_halt_poll_grow_start)
|
||||
val = 0;
|
||||
}
|
||||
|
||||
trace_guest_halt_poll_ns_shrink(val, dev->poll_limit_ns);
|
||||
dev->poll_limit_ns = val;
|
||||
}
|
||||
|
|
|
@ -44,6 +44,7 @@ static DEFINE_PER_CPU(struct ladder_device, ladder_devices);
|
|||
|
||||
/**
|
||||
* ladder_do_selection - prepares private data for a state change
|
||||
* @dev: the CPU
|
||||
* @ldev: the ladder device
|
||||
* @old_idx: the current state index
|
||||
* @new_idx: the new target state index
|
||||
|
|
|
@ -14,8 +14,6 @@
|
|||
#include <linux/ktime.h>
|
||||
#include <linux/hrtimer.h>
|
||||
#include <linux/tick.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/sched/loadavg.h>
|
||||
#include <linux/sched/stat.h>
|
||||
#include <linux/math64.h>
|
||||
|
||||
|
@ -95,16 +93,11 @@
|
|||
* state, and thus the less likely a busy CPU will hit such a deep
|
||||
* C state.
|
||||
*
|
||||
* Two factors are used in determing this multiplier:
|
||||
* a value of 10 is added for each point of "per cpu load average" we have.
|
||||
* a value of 5 points is added for each process that is waiting for
|
||||
* IO on this CPU.
|
||||
* (these values are experimentally determined)
|
||||
*
|
||||
* The load average factor gives a longer term (few seconds) input to the
|
||||
* decision, while the iowait value gives a cpu local instantanious input.
|
||||
* The iowait factor may look low, but realize that this is also already
|
||||
* represented in the system load average.
|
||||
* Currently there is only one value determining the factor:
|
||||
* 10 points are added for each process that is waiting for IO on this CPU.
|
||||
* (This value was experimentally determined.)
|
||||
* Utilization is no longer a factor as it was shown that it never contributed
|
||||
* significantly to the performance multiplier in the first place.
|
||||
*
|
||||
*/
|
||||
|
||||
|
|
|
@ -183,6 +183,23 @@ unsigned int cpumask_first_and(const struct cpumask *srcp1, const struct cpumask
|
|||
return find_first_and_bit(cpumask_bits(srcp1), cpumask_bits(srcp2), small_cpumask_bits);
|
||||
}
|
||||
|
||||
/**
|
||||
* cpumask_first_and_and - return the first cpu from *srcp1 & *srcp2 & *srcp3
|
||||
* @srcp1: the first input
|
||||
* @srcp2: the second input
|
||||
* @srcp3: the third input
|
||||
*
|
||||
* Return: >= nr_cpu_ids if no cpus set in all.
|
||||
*/
|
||||
static inline
|
||||
unsigned int cpumask_first_and_and(const struct cpumask *srcp1,
|
||||
const struct cpumask *srcp2,
|
||||
const struct cpumask *srcp3)
|
||||
{
|
||||
return find_first_and_and_bit(cpumask_bits(srcp1), cpumask_bits(srcp2),
|
||||
cpumask_bits(srcp3), small_cpumask_bits);
|
||||
}
|
||||
|
||||
/**
|
||||
* cpumask_last - get the last CPU in a cpumask
|
||||
* @srcp: - the cpumask pointer
|
||||
|
|
|
@ -29,6 +29,8 @@ unsigned long __find_nth_and_andnot_bit(const unsigned long *addr1, const unsign
|
|||
unsigned long n);
|
||||
extern unsigned long _find_first_and_bit(const unsigned long *addr1,
|
||||
const unsigned long *addr2, unsigned long size);
|
||||
unsigned long _find_first_and_and_bit(const unsigned long *addr1, const unsigned long *addr2,
|
||||
const unsigned long *addr3, unsigned long size);
|
||||
extern unsigned long _find_first_zero_bit(const unsigned long *addr, unsigned long size);
|
||||
extern unsigned long _find_last_bit(const unsigned long *addr, unsigned long size);
|
||||
|
||||
|
@ -345,6 +347,31 @@ unsigned long find_first_and_bit(const unsigned long *addr1,
|
|||
}
|
||||
#endif
|
||||
|
||||
/**
|
||||
* find_first_and_and_bit - find the first set bit in 3 memory regions
|
||||
* @addr1: The first address to base the search on
|
||||
* @addr2: The second address to base the search on
|
||||
* @addr3: The third address to base the search on
|
||||
* @size: The bitmap size in bits
|
||||
*
|
||||
* Returns the bit number for the first set bit
|
||||
* If no bits are set, returns @size.
|
||||
*/
|
||||
static inline
|
||||
unsigned long find_first_and_and_bit(const unsigned long *addr1,
|
||||
const unsigned long *addr2,
|
||||
const unsigned long *addr3,
|
||||
unsigned long size)
|
||||
{
|
||||
if (small_const_nbits(size)) {
|
||||
unsigned long val = *addr1 & *addr2 & *addr3 & GENMASK(size - 1, 0);
|
||||
|
||||
return val ? __ffs(val) : size;
|
||||
}
|
||||
|
||||
return _find_first_and_and_bit(addr1, addr2, addr3, size);
|
||||
}
|
||||
|
||||
#ifndef find_first_zero_bit
|
||||
/**
|
||||
* find_first_zero_bit - find the first cleared bit in a memory region
|
||||
|
|
|
@ -116,6 +116,18 @@ unsigned long _find_first_and_bit(const unsigned long *addr1,
|
|||
EXPORT_SYMBOL(_find_first_and_bit);
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Find the first set bit in three memory regions.
|
||||
*/
|
||||
unsigned long _find_first_and_and_bit(const unsigned long *addr1,
|
||||
const unsigned long *addr2,
|
||||
const unsigned long *addr3,
|
||||
unsigned long size)
|
||||
{
|
||||
return FIND_FIRST_BIT(addr1[idx] & addr2[idx] & addr3[idx], /* nop */, size);
|
||||
}
|
||||
EXPORT_SYMBOL(_find_first_and_and_bit);
|
||||
|
||||
#ifndef find_first_zero_bit
|
||||
/*
|
||||
* Find the first cleared bit in a memory region.
|
||||
|
|
Loading…
Reference in New Issue