1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
|
/* Copyright 2013-2014 IBM Corp.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
* implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/* This example code shows how to read from the trace buffer. */
#include <external/trace/trace.h>
#include "../ccan/endian/endian.h"
#include "../ccan/short_types/short_types.h"
#include <trace_types.h>
#include <errno.h>
bool trace_empty(const struct tracebuf *tb)
{
const struct trace_repeat *rep;
if (tb->rpos == tb->end)
return true;
/*
* If we have a single element only, and it's a repeat buffer
* we've already seen every repeat for (yet which may be
* incremented in future), we're also empty.
*/
rep = (void *)tb->buf + be64_to_cpu(tb->rpos & tb->mask);
if (be64_to_cpu(tb->end) != be64_to_cpu(tb->rpos) + sizeof(*rep))
return false;
if (rep->type != TRACE_REPEAT)
return false;
if (be16_to_cpu(rep->num) != be32_to_cpu(tb->last_repeat))
return false;
return true;
}
/* You can't read in parallel, so some locking required in caller. */
bool trace_get(union trace *t, struct tracebuf *tb)
{
u64 start, rpos;
size_t len;
len = sizeof(*t) < be32_to_cpu(tb->max_size) ? sizeof(*t) :
be32_to_cpu(tb->max_size);
if (trace_empty(tb))
return false;
again:
/*
* The actual buffer is slightly larger than tbsize, so this
* memcpy is always valid.
*/
memcpy(t, tb->buf + be64_to_cpu(tb->rpos & tb->mask), len);
rmb(); /* read barrier, so we read tb->start after copying record. */
start = be64_to_cpu(tb->start);
rpos = be64_to_cpu(tb->rpos);
/* Now, was that overwritten? */
if (rpos < start) {
/* Create overflow record. */
t->overflow.unused64 = 0;
t->overflow.type = TRACE_OVERFLOW;
t->overflow.len_div_8 = sizeof(t->overflow) / 8;
t->overflow.bytes_missed = cpu_to_be64(start - rpos);
tb->rpos = cpu_to_be64(start);
return true;
}
/* Repeat entries need special handling */
if (t->hdr.type == TRACE_REPEAT) {
u32 num = be16_to_cpu(t->repeat.num);
/* In case we've read some already... */
t->repeat.num = cpu_to_be16(num - be32_to_cpu(tb->last_repeat));
/* Record how many repeats we saw this time. */
tb->last_repeat = cpu_to_be32(num);
/* Don't report an empty repeat buffer. */
if (t->repeat.num == 0) {
/*
* This can't be the last buffer, otherwise
* trace_empty would have returned true.
*/
assert(be64_to_cpu(tb->end) >
rpos + t->hdr.len_div_8 * 8);
/* Skip to next entry. */
tb->rpos = cpu_to_be64(rpos + t->hdr.len_div_8 * 8);
tb->last_repeat = 0;
goto again;
}
} else {
tb->last_repeat = 0;
tb->rpos = cpu_to_be64(rpos + t->hdr.len_div_8 * 8);
}
return true;
}
|