aboutsummaryrefslogtreecommitdiff
path: root/trace
diff options
context:
space:
mode:
Diffstat (limited to 'trace')
-rw-r--r--trace/control-target.c9
-rw-r--r--trace/control.c21
-rw-r--r--trace/control.h3
-rw-r--r--trace/event-internal.h1
4 files changed, 23 insertions, 11 deletions
diff --git a/trace/control-target.c b/trace/control-target.c
index 74c029a..4ee3733 100644
--- a/trace/control-target.c
+++ b/trace/control-target.c
@@ -13,6 +13,15 @@
#include "translate-all.h"
+void trace_event_set_state_dynamic_init(TraceEvent *ev, bool state)
+{
+ TraceEventID id = trace_event_get_id(ev);
+ assert(trace_event_get_state_static(ev));
+ /* Ignore "vcpu" property, since no vCPUs have been created yet */
+ trace_events_enabled_count += state - trace_events_dstate[id];
+ trace_events_dstate[id] = state;
+}
+
void trace_event_set_state_dynamic(TraceEvent *ev, bool state)
{
CPUState *vcpu;
diff --git a/trace/control.c b/trace/control.c
index b179cde..05d85ac 100644
--- a/trace/control.c
+++ b/trace/control.c
@@ -34,8 +34,6 @@ int trace_events_enabled_count;
* - true : Integral counting the number of vCPUs that have this event enabled.
*/
uint16_t trace_events_dstate[TRACE_EVENT_COUNT];
-/* Marks events for late vCPU state init */
-static bool trace_events_dstate_init[TRACE_EVENT_COUNT];
QemuOptsList qemu_trace_opts = {
.name = "trace",
@@ -145,10 +143,7 @@ static void do_trace_enable_events(const char *line_buf)
TraceEvent *ev = NULL;
while ((ev = trace_event_pattern(line_ptr, ev)) != NULL) {
if (trace_event_get_state_static(ev)) {
- /* start tracing */
- trace_event_set_state_dynamic(ev, enable);
- /* mark for late vCPU init */
- trace_events_dstate_init[ev->id] = true;
+ trace_event_set_state_dynamic_init(ev, enable);
}
}
} else {
@@ -160,10 +155,7 @@ static void do_trace_enable_events(const char *line_buf)
error_report("WARNING: trace event '%s' is not traceable",
line_ptr);
} else {
- /* start tracing */
- trace_event_set_state_dynamic(ev, enable);
- /* mark for late vCPU init */
- trace_events_dstate_init[ev->id] = true;
+ trace_event_set_state_dynamic_init(ev, enable);
}
}
}
@@ -284,7 +276,14 @@ void trace_init_vcpu_events(void)
while ((ev = trace_event_pattern("*", ev)) != NULL) {
if (trace_event_is_vcpu(ev) &&
trace_event_get_state_static(ev) &&
- trace_events_dstate_init[ev->id]) {
+ trace_event_get_state_dynamic(ev)) {
+ TraceEventID id = trace_event_get_id(ev);
+ /* check preconditions */
+ assert(trace_events_dstate[id] == 1);
+ /* disable early-init state ... */
+ trace_events_dstate[id] = 0;
+ trace_events_enabled_count--;
+ /* ... and properly re-enable */
trace_event_set_state_dynamic(ev, true);
}
}
diff --git a/trace/control.h b/trace/control.h
index 0413b28..27a16fc 100644
--- a/trace/control.h
+++ b/trace/control.h
@@ -274,6 +274,9 @@ char *trace_opt_parse(const char *optarg);
*
* Re-synchronize initial event state with vCPUs (which can be created after
* trace_init_events()).
+ *
+ * Precondition: event states won't be changed between trace_enable_events() and
+ * trace_init_vcpu_events() (e.g., through QMP).
*/
void trace_init_vcpu_events(void);
diff --git a/trace/event-internal.h b/trace/event-internal.h
index 5d8fa97..074faf6 100644
--- a/trace/event-internal.h
+++ b/trace/event-internal.h
@@ -29,5 +29,6 @@ typedef struct TraceEvent {
const bool sstate;
} TraceEvent;
+void trace_event_set_state_dynamic_init(TraceEvent *ev, bool state);
#endif /* TRACE__EVENT_INTERNAL_H */