cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

cpu_pm.c (6288B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*
      3 * Copyright (C) 2011 Google, Inc.
      4 *
      5 * Author:
      6 *	Colin Cross <ccross@android.com>
      7 */
      8
      9#include <linux/kernel.h>
     10#include <linux/cpu_pm.h>
     11#include <linux/module.h>
     12#include <linux/notifier.h>
     13#include <linux/spinlock.h>
     14#include <linux/syscore_ops.h>
     15
     16/*
     17 * atomic_notifiers use a spinlock_t, which can block under PREEMPT_RT.
     18 * Notifications for cpu_pm will be issued by the idle task itself, which can
     19 * never block, IOW it requires using a raw_spinlock_t.
     20 */
     21static struct {
     22	struct raw_notifier_head chain;
     23	raw_spinlock_t lock;
     24} cpu_pm_notifier = {
     25	.chain = RAW_NOTIFIER_INIT(cpu_pm_notifier.chain),
     26	.lock  = __RAW_SPIN_LOCK_UNLOCKED(cpu_pm_notifier.lock),
     27};
     28
     29static int cpu_pm_notify(enum cpu_pm_event event)
     30{
     31	int ret;
     32
     33	/*
     34	 * This introduces a RCU read critical section, which could be
     35	 * disfunctional in cpu idle. Copy RCU_NONIDLE code to let RCU know
     36	 * this.
     37	 */
     38	rcu_irq_enter_irqson();
     39	rcu_read_lock();
     40	ret = raw_notifier_call_chain(&cpu_pm_notifier.chain, event, NULL);
     41	rcu_read_unlock();
     42	rcu_irq_exit_irqson();
     43
     44	return notifier_to_errno(ret);
     45}
     46
     47static int cpu_pm_notify_robust(enum cpu_pm_event event_up, enum cpu_pm_event event_down)
     48{
     49	unsigned long flags;
     50	int ret;
     51
     52	rcu_irq_enter_irqson();
     53	raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags);
     54	ret = raw_notifier_call_chain_robust(&cpu_pm_notifier.chain, event_up, event_down, NULL);
     55	raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags);
     56	rcu_irq_exit_irqson();
     57
     58	return notifier_to_errno(ret);
     59}
     60
     61/**
     62 * cpu_pm_register_notifier - register a driver with cpu_pm
     63 * @nb: notifier block to register
     64 *
     65 * Add a driver to a list of drivers that are notified about
     66 * CPU and CPU cluster low power entry and exit.
     67 *
     68 * This function has the same return conditions as raw_notifier_chain_register.
     69 */
     70int cpu_pm_register_notifier(struct notifier_block *nb)
     71{
     72	unsigned long flags;
     73	int ret;
     74
     75	raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags);
     76	ret = raw_notifier_chain_register(&cpu_pm_notifier.chain, nb);
     77	raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags);
     78	return ret;
     79}
     80EXPORT_SYMBOL_GPL(cpu_pm_register_notifier);
     81
     82/**
     83 * cpu_pm_unregister_notifier - unregister a driver with cpu_pm
     84 * @nb: notifier block to be unregistered
     85 *
     86 * Remove a driver from the CPU PM notifier list.
     87 *
     88 * This function has the same return conditions as raw_notifier_chain_unregister.
     89 */
     90int cpu_pm_unregister_notifier(struct notifier_block *nb)
     91{
     92	unsigned long flags;
     93	int ret;
     94
     95	raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags);
     96	ret = raw_notifier_chain_unregister(&cpu_pm_notifier.chain, nb);
     97	raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags);
     98	return ret;
     99}
    100EXPORT_SYMBOL_GPL(cpu_pm_unregister_notifier);
    101
    102/**
    103 * cpu_pm_enter - CPU low power entry notifier
    104 *
    105 * Notifies listeners that a single CPU is entering a low power state that may
    106 * cause some blocks in the same power domain as the cpu to reset.
    107 *
    108 * Must be called on the affected CPU with interrupts disabled.  Platform is
    109 * responsible for ensuring that cpu_pm_enter is not called twice on the same
    110 * CPU before cpu_pm_exit is called. Notified drivers can include VFP
    111 * co-processor, interrupt controller and its PM extensions, local CPU
    112 * timers context save/restore which shouldn't be interrupted. Hence it
    113 * must be called with interrupts disabled.
    114 *
    115 * Return conditions are same as __raw_notifier_call_chain.
    116 */
    117int cpu_pm_enter(void)
    118{
    119	return cpu_pm_notify_robust(CPU_PM_ENTER, CPU_PM_ENTER_FAILED);
    120}
    121EXPORT_SYMBOL_GPL(cpu_pm_enter);
    122
    123/**
    124 * cpu_pm_exit - CPU low power exit notifier
    125 *
    126 * Notifies listeners that a single CPU is exiting a low power state that may
    127 * have caused some blocks in the same power domain as the cpu to reset.
    128 *
    129 * Notified drivers can include VFP co-processor, interrupt controller
    130 * and its PM extensions, local CPU timers context save/restore which
    131 * shouldn't be interrupted. Hence it must be called with interrupts disabled.
    132 *
    133 * Return conditions are same as __raw_notifier_call_chain.
    134 */
    135int cpu_pm_exit(void)
    136{
    137	return cpu_pm_notify(CPU_PM_EXIT);
    138}
    139EXPORT_SYMBOL_GPL(cpu_pm_exit);
    140
    141/**
    142 * cpu_cluster_pm_enter - CPU cluster low power entry notifier
    143 *
    144 * Notifies listeners that all cpus in a power domain are entering a low power
    145 * state that may cause some blocks in the same power domain to reset.
    146 *
    147 * Must be called after cpu_pm_enter has been called on all cpus in the power
    148 * domain, and before cpu_pm_exit has been called on any cpu in the power
    149 * domain. Notified drivers can include VFP co-processor, interrupt controller
    150 * and its PM extensions, local CPU timers context save/restore which
    151 * shouldn't be interrupted. Hence it must be called with interrupts disabled.
    152 *
    153 * Must be called with interrupts disabled.
    154 *
    155 * Return conditions are same as __raw_notifier_call_chain.
    156 */
    157int cpu_cluster_pm_enter(void)
    158{
    159	return cpu_pm_notify_robust(CPU_CLUSTER_PM_ENTER, CPU_CLUSTER_PM_ENTER_FAILED);
    160}
    161EXPORT_SYMBOL_GPL(cpu_cluster_pm_enter);
    162
    163/**
    164 * cpu_cluster_pm_exit - CPU cluster low power exit notifier
    165 *
    166 * Notifies listeners that all cpus in a power domain are exiting form a
    167 * low power state that may have caused some blocks in the same power domain
    168 * to reset.
    169 *
    170 * Must be called after cpu_cluster_pm_enter has been called for the power
    171 * domain, and before cpu_pm_exit has been called on any cpu in the power
    172 * domain. Notified drivers can include VFP co-processor, interrupt controller
    173 * and its PM extensions, local CPU timers context save/restore which
    174 * shouldn't be interrupted. Hence it must be called with interrupts disabled.
    175 *
    176 * Return conditions are same as __raw_notifier_call_chain.
    177 */
    178int cpu_cluster_pm_exit(void)
    179{
    180	return cpu_pm_notify(CPU_CLUSTER_PM_EXIT);
    181}
    182EXPORT_SYMBOL_GPL(cpu_cluster_pm_exit);
    183
    184#ifdef CONFIG_PM
    185static int cpu_pm_suspend(void)
    186{
    187	int ret;
    188
    189	ret = cpu_pm_enter();
    190	if (ret)
    191		return ret;
    192
    193	ret = cpu_cluster_pm_enter();
    194	return ret;
    195}
    196
    197static void cpu_pm_resume(void)
    198{
    199	cpu_cluster_pm_exit();
    200	cpu_pm_exit();
    201}
    202
    203static struct syscore_ops cpu_pm_syscore_ops = {
    204	.suspend = cpu_pm_suspend,
    205	.resume = cpu_pm_resume,
    206};
    207
    208static int cpu_pm_init(void)
    209{
    210	register_syscore_ops(&cpu_pm_syscore_ops);
    211	return 0;
    212}
    213core_initcall(cpu_pm_init);
    214#endif