1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
|
/* Parts of target interface that deal with accessing memory and memory-like
objects.
Copyright (C) 2006-2018 Free Software Foundation, Inc.
This file is part of GDB.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>. */
#include "defs.h"
#include "vec.h"
#include "target.h"
#include "memory-map.h"
#include "gdb_sys_time.h"
#include <algorithm>
static bool
compare_block_starting_address (const memory_write_request &a_req,
const memory_write_request &b_req)
{
return a_req.begin < b_req.begin;
}
/* Adds to RESULT all memory write requests from BLOCK that are
in [BEGIN, END) range.
If any memory request is only partially in the specified range,
that part of the memory request will be added. */
static void
claim_memory (const std::vector<memory_write_request> &blocks,
std::vector<memory_write_request> *result,
ULONGEST begin,
ULONGEST end)
{
ULONGEST claimed_begin;
ULONGEST claimed_end;
for (const memory_write_request &r : blocks)
{
/* If the request doesn't overlap [BEGIN, END), skip it. We
must handle END == 0 meaning the top of memory; we don't yet
check for R->end == 0, which would also mean the top of
memory, but there's an assertion in
target_write_memory_blocks which checks for that. */
if (begin >= r.end)
continue;
if (end != 0 && end <= r.begin)
continue;
claimed_begin = std::max (begin, r.begin);
if (end == 0)
claimed_end = r.end;
else
claimed_end = std::min (end, r.end);
if (claimed_begin == r.begin && claimed_end == r.end)
result->push_back (r);
else
{
struct memory_write_request n = r;
n.begin = claimed_begin;
n.end = claimed_end;
n.data += claimed_begin - r.begin;
result->push_back (n);
}
}
}
/* Given a vector of struct memory_write_request objects in BLOCKS,
add memory requests for flash memory into FLASH_BLOCKS, and for
regular memory to REGULAR_BLOCKS. */
static void
split_regular_and_flash_blocks (const std::vector<memory_write_request> &blocks,
std::vector<memory_write_request> *regular_blocks,
std::vector<memory_write_request> *flash_blocks)
{
struct mem_region *region;
CORE_ADDR cur_address;
/* This implementation runs in O(length(regions)*length(blocks)) time.
However, in most cases the number of blocks will be small, so this does
not matter.
Note also that it's extremely unlikely that a memory write request
will span more than one memory region, however for safety we handle
such situations. */
cur_address = 0;
while (1)
{
std::vector<memory_write_request> *r;
region = lookup_mem_region (cur_address);
r = region->attrib.mode == MEM_FLASH ? flash_blocks : regular_blocks;
cur_address = region->hi;
claim_memory (blocks, r, region->lo, region->hi);
if (cur_address == 0)
break;
}
}
/* Given an ADDRESS, if BEGIN is non-NULL this function sets *BEGIN
to the start of the flash block containing the address. Similarly,
if END is non-NULL *END will be set to the address one past the end
of the block containing the address. */
static void
block_boundaries (CORE_ADDR address, CORE_ADDR *begin, CORE_ADDR *end)
{
struct mem_region *region;
unsigned blocksize;
CORE_ADDR offset_in_region;
region = lookup_mem_region (address);
gdb_assert (region->attrib.mode == MEM_FLASH);
blocksize = region->attrib.blocksize;
offset_in_region = address - region->lo;
if (begin)
*begin = region->lo + offset_in_region / blocksize * blocksize;
if (end)
*end = region->lo + (offset_in_region + blocksize - 1) / blocksize * blocksize;
}
/* Given the list of memory requests to be WRITTEN, this function
returns write requests covering each group of flash blocks which must
be erased. */
static std::vector<memory_write_request>
blocks_to_erase (const std::vector<memory_write_request> &written)
{
std::vector<memory_write_request> result;
for (const memory_write_request &request : written)
{
CORE_ADDR begin, end;
block_boundaries (request.begin, &begin, 0);
block_boundaries (request.end - 1, 0, &end);
if (!result.empty () && result.back ().end >= begin)
result.back ().end = end;
else
result.emplace_back (begin, end);
}
return result;
}
/* Given ERASED_BLOCKS, a list of blocks that will be erased with
flash erase commands, and WRITTEN_BLOCKS, the list of memory
addresses that will be written, compute the set of memory addresses
that will be erased but not rewritten (e.g. padding within a block
which is only partially filled by "load"). */
static std::vector<memory_write_request>
compute_garbled_blocks (const std::vector<memory_write_request> &erased_blocks,
const std::vector<memory_write_request> &written_blocks)
{
std::vector<memory_write_request> result;
unsigned j;
unsigned je = written_blocks.size ();
/* Look at each erased memory_write_request in turn, and
see what part of it is subsequently written to.
This implementation is O(length(erased) * length(written)). If
the lists are sorted at this point it could be rewritten more
efficiently, but the complexity is not generally worthwhile. */
for (const memory_write_request &erased_iter : erased_blocks)
{
/* Make a deep copy -- it will be modified inside the loop, but
we don't want to modify original vector. */
struct memory_write_request erased = erased_iter;
for (j = 0; j != je;)
{
const memory_write_request *written = &written_blocks[j];
/* Now try various cases. */
/* If WRITTEN is fully to the left of ERASED, check the next
written memory_write_request. */
if (written->end <= erased.begin)
{
++j;
continue;
}
/* If WRITTEN is fully to the right of ERASED, then ERASED
is not written at all. WRITTEN might affect other
blocks. */
if (written->begin >= erased.end)
{
result.push_back (erased);
goto next_erased;
}
/* If all of ERASED is completely written, we can move on to
the next erased region. */
if (written->begin <= erased.begin
&& written->end >= erased.end)
{
goto next_erased;
}
/* If there is an unwritten part at the beginning of ERASED,
then we should record that part and try this inner loop
again for the remainder. */
if (written->begin > erased.begin)
{
result.emplace_back (erased.begin, written->begin);
erased.begin = written->begin;
continue;
}
/* If there is an unwritten part at the end of ERASED, we
forget about the part that was written to and wait to see
if the next write request writes more of ERASED. We can't
push it yet. */
if (written->end < erased.end)
{
erased.begin = written->end;
++j;
continue;
}
}
/* If we ran out of write requests without doing anything about
ERASED, then that means it's really erased. */
result.push_back (erased);
next_erased:
;
}
return result;
}
int
target_write_memory_blocks (const std::vector<memory_write_request> &requests,
enum flash_preserve_mode preserve_flash_p,
void (*progress_cb) (ULONGEST, void *))
{
std::vector<memory_write_request> blocks = requests;
std::vector<memory_write_request> regular;
std::vector<memory_write_request> flash;
std::vector<memory_write_request> erased, garbled;
/* END == 0 would represent wraparound: a write to the very last
byte of the address space. This file was not written with that
possibility in mind. This is fixable, but a lot of work for a
rare problem; so for now, fail noisily here instead of obscurely
later. */
for (const memory_write_request &iter : requests)
gdb_assert (iter.end != 0);
/* Sort the blocks by their start address. */
std::sort (blocks.begin (), blocks.end (), compare_block_starting_address);
/* Split blocks into list of regular memory blocks,
and list of flash memory blocks. */
split_regular_and_flash_blocks (blocks, ®ular, &flash);
/* If a variable is added to forbid flash write, even during "load",
it should be checked here. Similarly, if this function is used
for other situations besides "load" in which writing to flash
is undesirable, that should be checked here. */
/* Find flash blocks to erase. */
erased = blocks_to_erase (flash);
/* Find what flash regions will be erased, and not overwritten; then
either preserve or discard the old contents. */
garbled = compute_garbled_blocks (erased, flash);
std::vector<gdb::unique_xmalloc_ptr<gdb_byte>> mem_holders;
if (!garbled.empty ())
{
if (preserve_flash_p == flash_preserve)
{
/* Read in regions that must be preserved and add them to
the list of blocks we read. */
for (memory_write_request &iter : garbled)
{
gdb_assert (iter.data == NULL);
gdb::unique_xmalloc_ptr<gdb_byte> holder
((gdb_byte *) xmalloc (iter.end - iter.begin));
iter.data = holder.get ();
mem_holders.push_back (std::move (holder));
int err = target_read_memory (iter.begin, iter.data,
iter.end - iter.begin);
if (err != 0)
return err;
flash.push_back (iter);
}
std::sort (flash.begin (), flash.end (),
compare_block_starting_address);
}
}
/* We could coalesce adjacent memory blocks here, to reduce the
number of write requests for small sections. However, we would
have to reallocate and copy the data pointers, which could be
large; large sections are more common in loadable objects than
large numbers of small sections (although the reverse can be true
in object files). So, we issue at least one write request per
passed struct memory_write_request. The remote stub will still
have the opportunity to batch flash requests. */
/* Write regular blocks. */
for (const memory_write_request &iter : regular)
{
LONGEST len;
len = target_write_with_progress (current_top_target (),
TARGET_OBJECT_MEMORY, NULL,
iter.data, iter.begin,
iter.end - iter.begin,
progress_cb, iter.baton);
if (len < (LONGEST) (iter.end - iter.begin))
{
/* Call error? */
return -1;
}
}
if (!erased.empty ())
{
/* Erase all pages. */
for (const memory_write_request &iter : erased)
target_flash_erase (iter.begin, iter.end - iter.begin);
/* Write flash data. */
for (const memory_write_request &iter : flash)
{
LONGEST len;
len = target_write_with_progress (current_top_target (),
TARGET_OBJECT_FLASH, NULL,
iter.data, iter.begin,
iter.end - iter.begin,
progress_cb, iter.baton);
if (len < (LONGEST) (iter.end - iter.begin))
error (_("Error writing data to flash"));
}
target_flash_done ();
}
return 0;
}
|