diff options
author | Tejun Heo <tj@kernel.org> | 2024-02-20 19:36:13 -1000 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2024-02-20 19:36:13 -1000 |
commit | c7a40c49af920fbad2ab6795b6587308ad69de9f (patch) | |
tree | 11c9b122ddc2583483f2e8249198dea0e4110cbe /kernel/workqueue.c | |
parent | fd0a68a2337b79a7bd4dad5e7d9dc726828527af (diff) | |
download | linux-c7a40c49af920fbad2ab6795b6587308ad69de9f.tar.gz linux-c7a40c49af920fbad2ab6795b6587308ad69de9f.tar.bz2 linux-c7a40c49af920fbad2ab6795b6587308ad69de9f.zip |
workqueue: Cosmetic changes
Reorder some global declarations and adjust comments and whitespaces for
clarity and consistency. No functional changes.
Signed-off-by: Tejun Heo <tj@kernel.org>
Reviewed-by: Lai Jiangshan <jiangshanlai@gmail.com>
Diffstat (limited to 'kernel/workqueue.c')
-rw-r--r-- | kernel/workqueue.c | 30 |
1 files changed, 14 insertions, 16 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 6ae441e13804..b280caf81fb2 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -376,8 +376,6 @@ struct workqueue_struct { struct wq_node_nr_active *node_nr_active[]; /* I: per-node nr_active */ }; -static struct kmem_cache *pwq_cache; - /* * Each pod type describes how CPUs should be grouped for unbound workqueues. * See the comment above workqueue_attrs->affn_scope. @@ -389,20 +387,15 @@ struct wq_pod_type { int *cpu_pod; /* cpu -> pod */ }; -static struct wq_pod_type wq_pod_types[WQ_AFFN_NR_TYPES]; -static enum wq_affn_scope wq_affn_dfl = WQ_AFFN_CACHE; - static const char *wq_affn_names[WQ_AFFN_NR_TYPES] = { - [WQ_AFFN_DFL] = "default", - [WQ_AFFN_CPU] = "cpu", - [WQ_AFFN_SMT] = "smt", - [WQ_AFFN_CACHE] = "cache", - [WQ_AFFN_NUMA] = "numa", - [WQ_AFFN_SYSTEM] = "system", + [WQ_AFFN_DFL] = "default", + [WQ_AFFN_CPU] = "cpu", + [WQ_AFFN_SMT] = "smt", + [WQ_AFFN_CACHE] = "cache", + [WQ_AFFN_NUMA] = "numa", + [WQ_AFFN_SYSTEM] = "system", }; -static bool wq_topo_initialized __read_mostly = false; - /* * Per-cpu work items which run for longer than the following threshold are * automatically considered CPU intensive and excluded from concurrency @@ -418,6 +411,12 @@ static bool wq_power_efficient = IS_ENABLED(CONFIG_WQ_POWER_EFFICIENT_DEFAULT); module_param_named(power_efficient, wq_power_efficient, bool, 0444); static bool wq_online; /* can kworkers be created yet? */ +static bool wq_topo_initialized __read_mostly = false; + +static struct kmem_cache *pwq_cache; + +static struct wq_pod_type wq_pod_types[WQ_AFFN_NR_TYPES]; +static enum wq_affn_scope wq_affn_dfl = WQ_AFFN_CACHE; /* buf for wq_update_unbound_pod_attrs(), protected by CPU hotplug exclusion */ static struct workqueue_attrs *wq_update_pod_attrs_buf; @@ -2231,7 +2230,6 @@ static void __queue_work(int cpu, struct workqueue_struct *wq, */ lockdep_assert_irqs_disabled(); - /* * For a draining wq, only works from the same workqueue are * allowed. The __WQ_DESTROYING helps to spot the issue that @@ -4121,8 +4119,8 @@ static bool __cancel_work_timer(struct work_struct *work, bool is_dwork) local_irq_restore(flags); /* - * This allows canceling during early boot. We know that @work - * isn't executing. + * Skip __flush_work() during early boot when we know that @work isn't + * executing. This allows canceling during early boot. */ if (wq_online) __flush_work(work, true); |