LCOV - code coverage report
Current view: top level - kernel/rcu - srcutiny.c (source / functions) Hit Total Coverage
Test: coverage.info Lines: 73 101 72.3 %
Date: 2023-07-19 18:55:55 Functions: 9 13 69.2 %

          Line data    Source code
       1             : // SPDX-License-Identifier: GPL-2.0+
       2             : /*
       3             :  * Sleepable Read-Copy Update mechanism for mutual exclusion,
       4             :  *      tiny version for non-preemptible single-CPU use.
       5             :  *
       6             :  * Copyright (C) IBM Corporation, 2017
       7             :  *
       8             :  * Author: Paul McKenney <paulmck@linux.ibm.com>
       9             :  */
      10             : 
      11             : #include <linux/export.h>
      12             : #include <linux/mutex.h>
      13             : #include <linux/preempt.h>
      14             : #include <linux/rcupdate_wait.h>
      15             : #include <linux/sched.h>
      16             : #include <linux/delay.h>
      17             : #include <linux/srcu.h>
      18             : 
      19             : #include <linux/rcu_node_tree.h>
      20             : #include "rcu_segcblist.h"
      21             : #include "rcu.h"
      22             : 
      23             : int rcu_scheduler_active __read_mostly;
      24             : static LIST_HEAD(srcu_boot_list);
      25             : static bool srcu_init_done;
      26             : 
      27           2 : static int init_srcu_struct_fields(struct srcu_struct *ssp)
      28             : {
      29           2 :         ssp->srcu_lock_nesting[0] = 0;
      30           2 :         ssp->srcu_lock_nesting[1] = 0;
      31           2 :         init_swait_queue_head(&ssp->srcu_wq);
      32           2 :         ssp->srcu_cb_head = NULL;
      33           2 :         ssp->srcu_cb_tail = &ssp->srcu_cb_head;
      34           2 :         ssp->srcu_gp_running = false;
      35           2 :         ssp->srcu_gp_waiting = false;
      36           2 :         ssp->srcu_idx = 0;
      37           2 :         ssp->srcu_idx_max = 0;
      38           2 :         INIT_WORK(&ssp->srcu_work, srcu_drive_gp);
      39           4 :         INIT_LIST_HEAD(&ssp->srcu_work.entry);
      40           2 :         return 0;
      41             : }
      42             : 
      43             : #ifdef CONFIG_DEBUG_LOCK_ALLOC
      44             : 
      45             : int __init_srcu_struct(struct srcu_struct *ssp, const char *name,
      46             :                        struct lock_class_key *key)
      47             : {
      48             :         /* Don't re-initialize a lock while it is held. */
      49             :         debug_check_no_locks_freed((void *)ssp, sizeof(*ssp));
      50             :         lockdep_init_map(&ssp->dep_map, name, key, 0);
      51             :         return init_srcu_struct_fields(ssp);
      52             : }
      53             : EXPORT_SYMBOL_GPL(__init_srcu_struct);
      54             : 
      55             : #else /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */
      56             : 
      57             : /*
      58             :  * init_srcu_struct - initialize a sleep-RCU structure
      59             :  * @ssp: structure to initialize.
      60             :  *
      61             :  * Must invoke this on a given srcu_struct before passing that srcu_struct
      62             :  * to any other function.  Each srcu_struct represents a separate domain
      63             :  * of SRCU protection.
      64             :  */
      65           2 : int init_srcu_struct(struct srcu_struct *ssp)
      66             : {
      67           2 :         return init_srcu_struct_fields(ssp);
      68             : }
      69             : EXPORT_SYMBOL_GPL(init_srcu_struct);
      70             : 
      71             : #endif /* #else #ifdef CONFIG_DEBUG_LOCK_ALLOC */
      72             : 
      73             : /*
      74             :  * cleanup_srcu_struct - deconstruct a sleep-RCU structure
      75             :  * @ssp: structure to clean up.
      76             :  *
      77             :  * Must invoke this after you are finished using a given srcu_struct that
      78             :  * was initialized via init_srcu_struct(), else you leak memory.
      79             :  */
      80           0 : void cleanup_srcu_struct(struct srcu_struct *ssp)
      81             : {
      82           0 :         WARN_ON(ssp->srcu_lock_nesting[0] || ssp->srcu_lock_nesting[1]);
      83           0 :         flush_work(&ssp->srcu_work);
      84           0 :         WARN_ON(ssp->srcu_gp_running);
      85           0 :         WARN_ON(ssp->srcu_gp_waiting);
      86           0 :         WARN_ON(ssp->srcu_cb_head);
      87           0 :         WARN_ON(&ssp->srcu_cb_head != ssp->srcu_cb_tail);
      88           0 :         WARN_ON(ssp->srcu_idx != ssp->srcu_idx_max);
      89           0 :         WARN_ON(ssp->srcu_idx & 0x1);
      90           0 : }
      91             : EXPORT_SYMBOL_GPL(cleanup_srcu_struct);
      92             : 
      93             : /*
      94             :  * Removes the count for the old reader from the appropriate element of
      95             :  * the srcu_struct.
      96             :  */
      97        1997 : void __srcu_read_unlock(struct srcu_struct *ssp, int idx)
      98             : {
      99        1997 :         int newval = READ_ONCE(ssp->srcu_lock_nesting[idx]) - 1;
     100             : 
     101        1997 :         WRITE_ONCE(ssp->srcu_lock_nesting[idx], newval);
     102        1997 :         if (!newval && READ_ONCE(ssp->srcu_gp_waiting) && in_task())
     103           0 :                 swake_up_one(&ssp->srcu_wq);
     104        1997 : }
     105             : EXPORT_SYMBOL_GPL(__srcu_read_unlock);
     106             : 
     107             : /*
     108             :  * Workqueue handler to drive one grace period and invoke any callbacks
     109             :  * that become ready as a result.  Single-CPU and !PREEMPTION operation
     110             :  * means that we get away with murder on synchronization.  ;-)
     111             :  */
     112          22 : void srcu_drive_gp(struct work_struct *wp)
     113             : {
     114             :         int idx;
     115             :         struct rcu_head *lh;
     116             :         struct rcu_head *rhp;
     117             :         struct srcu_struct *ssp;
     118             : 
     119          22 :         ssp = container_of(wp, struct srcu_struct, srcu_work);
     120          22 :         if (ssp->srcu_gp_running || ULONG_CMP_GE(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max)))
     121             :                 return; /* Already running or nothing to do. */
     122             : 
     123             :         /* Remove recently arrived callbacks and wait for readers. */
     124          22 :         WRITE_ONCE(ssp->srcu_gp_running, true);
     125             :         local_irq_disable();
     126          22 :         lh = ssp->srcu_cb_head;
     127          22 :         ssp->srcu_cb_head = NULL;
     128          22 :         ssp->srcu_cb_tail = &ssp->srcu_cb_head;
     129             :         local_irq_enable();
     130          22 :         idx = (ssp->srcu_idx & 0x2) / 2;
     131          22 :         WRITE_ONCE(ssp->srcu_idx, ssp->srcu_idx + 1);
     132          22 :         WRITE_ONCE(ssp->srcu_gp_waiting, true);  /* srcu_read_unlock() wakes! */
     133          22 :         swait_event_exclusive(ssp->srcu_wq, !READ_ONCE(ssp->srcu_lock_nesting[idx]));
     134          22 :         WRITE_ONCE(ssp->srcu_gp_waiting, false); /* srcu_read_unlock() cheap. */
     135          22 :         WRITE_ONCE(ssp->srcu_idx, ssp->srcu_idx + 1);
     136             : 
     137             :         /* Invoke the callbacks we removed above. */
     138          66 :         while (lh) {
     139          22 :                 rhp = lh;
     140          22 :                 lh = lh->next;
     141          22 :                 local_bh_disable();
     142          22 :                 rhp->func(rhp);
     143             :                 local_bh_enable();
     144             :         }
     145             : 
     146             :         /*
     147             :          * Enable rescheduling, and if there are more callbacks,
     148             :          * reschedule ourselves.  This can race with a call_srcu()
     149             :          * at interrupt level, but the ->srcu_gp_running checks will
     150             :          * straighten that out.
     151             :          */
     152          22 :         WRITE_ONCE(ssp->srcu_gp_running, false);
     153          22 :         if (ULONG_CMP_LT(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max)))
     154           0 :                 schedule_work(&ssp->srcu_work);
     155             : }
     156             : EXPORT_SYMBOL_GPL(srcu_drive_gp);
     157             : 
     158          22 : static void srcu_gp_start_if_needed(struct srcu_struct *ssp)
     159             : {
     160             :         unsigned long cookie;
     161             : 
     162          22 :         cookie = get_state_synchronize_srcu(ssp);
     163          22 :         if (ULONG_CMP_GE(READ_ONCE(ssp->srcu_idx_max), cookie))
     164             :                 return;
     165          22 :         WRITE_ONCE(ssp->srcu_idx_max, cookie);
     166          22 :         if (!READ_ONCE(ssp->srcu_gp_running)) {
     167          22 :                 if (likely(srcu_init_done))
     168          22 :                         schedule_work(&ssp->srcu_work);
     169           0 :                 else if (list_empty(&ssp->srcu_work.entry))
     170           0 :                         list_add(&ssp->srcu_work.entry, &srcu_boot_list);
     171             :         }
     172             : }
     173             : 
     174             : /*
     175             :  * Enqueue an SRCU callback on the specified srcu_struct structure,
     176             :  * initiating grace-period processing if it is not already running.
     177             :  */
     178          22 : void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp,
     179             :                rcu_callback_t func)
     180             : {
     181             :         unsigned long flags;
     182             : 
     183          22 :         rhp->func = func;
     184          22 :         rhp->next = NULL;
     185          22 :         local_irq_save(flags);
     186          22 :         *ssp->srcu_cb_tail = rhp;
     187          22 :         ssp->srcu_cb_tail = &rhp->next;
     188          44 :         local_irq_restore(flags);
     189          22 :         srcu_gp_start_if_needed(ssp);
     190          22 : }
     191             : EXPORT_SYMBOL_GPL(call_srcu);
     192             : 
     193             : /*
     194             :  * synchronize_srcu - wait for prior SRCU read-side critical-section completion
     195             :  */
     196          22 : void synchronize_srcu(struct srcu_struct *ssp)
     197             : {
     198             :         struct rcu_synchronize rs;
     199             : 
     200             :         srcu_lock_sync(&ssp->dep_map);
     201             : 
     202             :         RCU_LOCKDEP_WARN(lockdep_is_held(ssp) ||
     203             :                         lock_is_held(&rcu_bh_lock_map) ||
     204             :                         lock_is_held(&rcu_lock_map) ||
     205             :                         lock_is_held(&rcu_sched_lock_map),
     206             :                         "Illegal synchronize_srcu() in same-type SRCU (or in RCU) read-side critical section");
     207             : 
     208          22 :         if (rcu_scheduler_active == RCU_SCHEDULER_INACTIVE)
     209           0 :                 return;
     210             : 
     211             :         might_sleep();
     212          22 :         init_rcu_head_on_stack(&rs.head);
     213          22 :         init_completion(&rs.completion);
     214          22 :         call_srcu(ssp, &rs.head, wakeme_after_rcu);
     215          22 :         wait_for_completion(&rs.completion);
     216          22 :         destroy_rcu_head_on_stack(&rs.head);
     217             : }
     218             : EXPORT_SYMBOL_GPL(synchronize_srcu);
     219             : 
     220             : /*
     221             :  * get_state_synchronize_srcu - Provide an end-of-grace-period cookie
     222             :  */
     223           0 : unsigned long get_state_synchronize_srcu(struct srcu_struct *ssp)
     224             : {
     225             :         unsigned long ret;
     226             : 
     227          22 :         barrier();
     228          22 :         ret = (READ_ONCE(ssp->srcu_idx) + 3) & ~0x1;
     229          22 :         barrier();
     230           0 :         return ret;
     231             : }
     232             : EXPORT_SYMBOL_GPL(get_state_synchronize_srcu);
     233             : 
     234             : /*
     235             :  * start_poll_synchronize_srcu - Provide cookie and start grace period
     236             :  *
     237             :  * The difference between this and get_state_synchronize_srcu() is that
     238             :  * this function ensures that the poll_state_synchronize_srcu() will
     239             :  * eventually return the value true.
     240             :  */
     241           0 : unsigned long start_poll_synchronize_srcu(struct srcu_struct *ssp)
     242             : {
     243           0 :         unsigned long ret = get_state_synchronize_srcu(ssp);
     244             : 
     245           0 :         srcu_gp_start_if_needed(ssp);
     246           0 :         return ret;
     247             : }
     248             : EXPORT_SYMBOL_GPL(start_poll_synchronize_srcu);
     249             : 
     250             : /*
     251             :  * poll_state_synchronize_srcu - Has cookie's grace period ended?
     252             :  */
     253           0 : bool poll_state_synchronize_srcu(struct srcu_struct *ssp, unsigned long cookie)
     254             : {
     255           0 :         unsigned long cur_s = READ_ONCE(ssp->srcu_idx);
     256             : 
     257           0 :         barrier();
     258           0 :         return ULONG_CMP_GE(cur_s, cookie) || ULONG_CMP_LT(cur_s, cookie - 3);
     259             : }
     260             : EXPORT_SYMBOL_GPL(poll_state_synchronize_srcu);
     261             : 
     262             : /* Lockdep diagnostics.  */
     263           1 : void __init rcu_scheduler_starting(void)
     264             : {
     265           1 :         rcu_scheduler_active = RCU_SCHEDULER_RUNNING;
     266           1 : }
     267             : 
     268             : /*
     269             :  * Queue work for srcu_struct structures with early boot callbacks.
     270             :  * The work won't actually execute until the workqueue initialization
     271             :  * phase that takes place after the scheduler starts.
     272             :  */
     273           1 : void __init srcu_init(void)
     274             : {
     275             :         struct srcu_struct *ssp;
     276             : 
     277           1 :         srcu_init_done = true;
     278           2 :         while (!list_empty(&srcu_boot_list)) {
     279           0 :                 ssp = list_first_entry(&srcu_boot_list,
     280             :                                       struct srcu_struct, srcu_work.entry);
     281           0 :                 list_del_init(&ssp->srcu_work.entry);
     282           0 :                 schedule_work(&ssp->srcu_work);
     283             :         }
     284           1 : }

Generated by: LCOV version 1.14