summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorValentin Schneider <vschneid@redhat.com>2023-03-07 14:35:55 +0000
committerPeter Zijlstra <peterz@infradead.org>2023-03-24 11:01:27 +0100
commit4468161a5ca2ea239c92de7c0a0dca61854ec4da (patch)
tree70a0aae983c49bc62519004eb4261a910d870be0
parent08407b5f61c1bbd4ebb26a76474df4354fd76fb7 (diff)
downloadlinux-4468161a5ca2ea239c92de7c0a0dca61854ec4da.tar.gz
linux-4468161a5ca2ea239c92de7c0a0dca61854ec4da.tar.bz2
linux-4468161a5ca2ea239c92de7c0a0dca61854ec4da.zip
irq_work: Trace self-IPIs sent via arch_irq_work_raise()
IPIs sent to remote CPUs via irq_work_queue_on() are now covered by trace_ipi_send_cpumask(), add another instance of the tracepoint to cover self-IPIs. Signed-off-by: Valentin Schneider <vschneid@redhat.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Steven Rostedt (Google) <rostedt@goodmis.org> Link: https://lore.kernel.org/r/20230307143558.294354-5-vschneid@redhat.com
-rw-r--r--kernel/irq_work.c14
1 files changed, 13 insertions, 1 deletions
diff --git a/kernel/irq_work.c b/kernel/irq_work.c
index 7afa40fe5cc4..c33e88e32a67 100644
--- a/kernel/irq_work.c
+++ b/kernel/irq_work.c
@@ -22,6 +22,8 @@
#include <asm/processor.h>
#include <linux/kasan.h>
+#include <trace/events/ipi.h>
+
static DEFINE_PER_CPU(struct llist_head, raised_list);
static DEFINE_PER_CPU(struct llist_head, lazy_list);
static DEFINE_PER_CPU(struct task_struct *, irq_workd);
@@ -74,6 +76,16 @@ void __weak arch_irq_work_raise(void)
*/
}
+static __always_inline void irq_work_raise(struct irq_work *work)
+{
+ if (trace_ipi_send_cpumask_enabled() && arch_irq_work_has_interrupt())
+ trace_ipi_send_cpumask(cpumask_of(smp_processor_id()),
+ _RET_IP_,
+ work->func);
+
+ arch_irq_work_raise();
+}
+
/* Enqueue on current CPU, work must already be claimed and preempt disabled */
static void __irq_work_queue_local(struct irq_work *work)
{
@@ -99,7 +111,7 @@ static void __irq_work_queue_local(struct irq_work *work)
/* If the work is "lazy", handle it from next tick if any */
if (!lazy_work || tick_nohz_tick_stopped())
- arch_irq_work_raise();
+ irq_work_raise(work);
}
/* Enqueue the irq work @work on the current CPU */