diff options
author | Markus Metzger <markus.t.metzger@intel.com> | 2014-01-17 14:40:02 +0100 |
---|---|---|
committer | Markus Metzger <markus.t.metzger@intel.com> | 2015-02-09 09:33:59 +0100 |
commit | aadf7753fd4cc3d9eb1cd0c089fd7a483b58f59e (patch) | |
tree | c2ecfeb8490e4595b14c2a3ae89c6a094b51f3ab /gdb/nat/linux-btrace.c | |
parent | 043c35779713a14e0916a1b3e31e006cd1270ee4 (diff) | |
download | gdb-aadf7753fd4cc3d9eb1cd0c089fd7a483b58f59e.zip gdb-aadf7753fd4cc3d9eb1cd0c089fd7a483b58f59e.tar.gz gdb-aadf7753fd4cc3d9eb1cd0c089fd7a483b58f59e.tar.bz2 |
btrace, linux: add perf event buffer abstraction
Collect perf event buffer related fields from btrace_target_info into
a new struct perf_event_buffer. Update functions that operated on the
buffer to take a struct perf_event_buffer pointer rather than a
btrace_target_info pointer.
2015-02-09 Markus Metzger <markus.t.metzger@intel.com>
* nat/linux-btrace.h (perf_event_buffer): New.
(btrace_target_info) <buffer, size, data_head>: Replace with ...
<bts>: ... this.
* nat/linux-btrace.c (perf_event_header, perf_event_mmap_size)
(perf_event_buffer_size, perf_event_buffer_begin)
(perf_event_buffer_end, linux_btrace_has_changed): Removed.
Updated users.
(perf_event_new_data): New.
Diffstat (limited to 'gdb/nat/linux-btrace.c')
-rw-r--r-- | gdb/nat/linux-btrace.c | 113 |
1 files changed, 38 insertions, 75 deletions
diff --git a/gdb/nat/linux-btrace.c b/gdb/nat/linux-btrace.c index 08eb49b..5b22661 100644 --- a/gdb/nat/linux-btrace.c +++ b/gdb/nat/linux-btrace.c @@ -59,45 +59,12 @@ struct perf_event_sample struct perf_event_bts bts; }; -/* Get the perf_event header. */ +/* Return non-zero if there is new data in PEVENT; zero otherwise. */ -static inline volatile struct perf_event_mmap_page * -perf_event_header (struct btrace_target_info* tinfo) -{ - return tinfo->buffer; -} - -/* Get the size of the perf_event mmap buffer. */ - -static inline size_t -perf_event_mmap_size (const struct btrace_target_info *tinfo) -{ - /* The branch trace buffer is preceded by a configuration page. */ - return (tinfo->size + 1) * PAGE_SIZE; -} - -/* Get the size of the perf_event buffer. */ - -static inline size_t -perf_event_buffer_size (struct btrace_target_info* tinfo) -{ - return tinfo->size * PAGE_SIZE; -} - -/* Get the start address of the perf_event buffer. */ - -static inline const uint8_t * -perf_event_buffer_begin (struct btrace_target_info* tinfo) -{ - return ((const uint8_t *) tinfo->buffer) + PAGE_SIZE; -} - -/* Get the end address of the perf_event buffer. */ - -static inline const uint8_t * -perf_event_buffer_end (struct btrace_target_info* tinfo) +static int +perf_event_new_data (const struct perf_event_buffer *pev) { - return perf_event_buffer_begin (tinfo) + perf_event_buffer_size (tinfo); + return *pev->data_head != pev->last_head; } /* Check whether an address is in the kernel. */ @@ -162,11 +129,12 @@ perf_event_sample_ok (const struct perf_event_sample *sample) static VEC (btrace_block_s) * perf_event_read_bts (struct btrace_target_info* tinfo, const uint8_t *begin, - const uint8_t *end, const uint8_t *start, size_t size) + const uint8_t *end, const uint8_t *start, + unsigned long long size) { VEC (btrace_block_s) *btrace = NULL; struct perf_event_sample sample; - size_t read = 0; + unsigned long long read = 0; struct btrace_block block = { 0, 0 }; struct regcache *regcache; @@ -432,6 +400,7 @@ linux_supports_btrace (struct target_ops *ops, enum btrace_format format) struct btrace_target_info * linux_enable_btrace (ptid_t ptid) { + struct perf_event_mmap_page *header; struct btrace_target_info *tinfo; int pid, pg; @@ -467,15 +436,24 @@ linux_enable_btrace (ptid_t ptid) for (pg = 4; pg >= 0; --pg) { /* The number of pages we request needs to be a power of two. */ - tinfo->size = 1 << pg; - tinfo->buffer = mmap (NULL, perf_event_mmap_size (tinfo), - PROT_READ, MAP_SHARED, tinfo->file, 0); - if (tinfo->buffer == MAP_FAILED) - continue; - - return tinfo; + header = mmap (NULL, ((1 << pg) + 1) * PAGE_SIZE, PROT_READ, MAP_SHARED, + tinfo->file, 0); + if (header != MAP_FAILED) + break; } + if (header == MAP_FAILED) + goto err_file; + + tinfo->header = header; + tinfo->bts.mem = ((const uint8_t *) header) + PAGE_SIZE; + tinfo->bts.size = (1 << pg) * PAGE_SIZE; + tinfo->bts.data_head = &header->data_head; + tinfo->bts.last_head = 0; + + return tinfo; + + err_file: /* We were not able to allocate any buffer. */ close (tinfo->file); @@ -489,29 +467,13 @@ linux_enable_btrace (ptid_t ptid) enum btrace_error linux_disable_btrace (struct btrace_target_info *tinfo) { - int errcode; - - errno = 0; - errcode = munmap (tinfo->buffer, perf_event_mmap_size (tinfo)); - if (errcode != 0) - return BTRACE_ERR_UNKNOWN; - + munmap((void *) tinfo->header, tinfo->bts.size + PAGE_SIZE); close (tinfo->file); xfree (tinfo); return BTRACE_ERR_NONE; } -/* Check whether the branch trace has changed. */ - -static int -linux_btrace_has_changed (struct btrace_target_info *tinfo) -{ - volatile struct perf_event_mmap_page *header = perf_event_header (tinfo); - - return header->data_head != tinfo->data_head; -} - /* Read branch trace data in BTS format for the thread given by TINFO into BTRACE using the TYPE reading method. */ @@ -520,24 +482,25 @@ linux_read_bts (struct btrace_data_bts *btrace, struct btrace_target_info *tinfo, enum btrace_read_type type) { - volatile struct perf_event_mmap_page *header; + struct perf_event_buffer *pevent; const uint8_t *begin, *end, *start; - unsigned long data_head, data_tail, retries = 5; - size_t buffer_size, size; + unsigned long long data_head, data_tail, buffer_size, size; + unsigned int retries = 5; + + pevent = &tinfo->bts; /* For delta reads, we return at least the partial last block containing the current PC. */ - if (type == BTRACE_READ_NEW && !linux_btrace_has_changed (tinfo)) + if (type == BTRACE_READ_NEW && !perf_event_new_data (pevent)) return BTRACE_ERR_NONE; - header = perf_event_header (tinfo); - buffer_size = perf_event_buffer_size (tinfo); - data_tail = tinfo->data_head; + buffer_size = pevent->size; + data_tail = pevent->last_head; /* We may need to retry reading the trace. See below. */ while (retries--) { - data_head = header->data_head; + data_head = *pevent->data_head; /* Delete any leftover trace from the previous iteration. */ VEC_free (btrace_block_s, btrace->blocks); @@ -569,13 +532,13 @@ linux_read_bts (struct btrace_data_bts *btrace, } /* Data_head keeps growing; the buffer itself is circular. */ - begin = perf_event_buffer_begin (tinfo); + begin = pevent->mem; start = begin + data_head % buffer_size; if (data_head <= buffer_size) end = start; else - end = perf_event_buffer_end (tinfo); + end = begin + pevent->size; btrace->blocks = perf_event_read_bts (tinfo, begin, end, start, size); @@ -584,11 +547,11 @@ linux_read_bts (struct btrace_data_bts *btrace, kernel might be writing the last branch trace records. Let's check whether the data head moved while we read the trace. */ - if (data_head == header->data_head) + if (data_head == *pevent->data_head) break; } - tinfo->data_head = data_head; + pevent->last_head = data_head; /* Prune the incomplete last block (i.e. the first one of inferior execution) if we're not doing a delta read. There is no way of filling in its zeroed |