1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
|
/* Ada Ravenscar thread support.
Copyright (C) 2004-2023 Free Software Foundation, Inc.
This file is part of GDB.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>. */
#include "defs.h"
#include "gdbcore.h"
#include "gdbthread.h"
#include "ada-lang.h"
#include "target.h"
#include "inferior.h"
#include "command.h"
#include "ravenscar-thread.h"
#include "observable.h"
#include "gdbcmd.h"
#include "top.h"
#include "regcache.h"
#include "objfiles.h"
#include <unordered_map>
/* This module provides support for "Ravenscar" tasks (Ada) when
debugging on bare-metal targets.
The typical situation is when debugging a bare-metal target over
the remote protocol. In that situation, the system does not know
about high-level concepts such as threads, only about some code
running on one or more CPUs. And since the remote protocol does not
provide any handling for CPUs, the de facto standard for handling
them is to have one thread per CPU, where the thread's ptid has
its lwp field set to the CPU number (eg: 1 for the first CPU,
2 for the second one, etc). This module will make that assumption.
This module then creates and maintains the list of threads based
on the list of Ada tasks, with one thread per Ada task. The convention
is that threads corresponding to the CPUs (see assumption above)
have a ptid_t of the form (PID, LWP, 0), while threads corresponding
to our Ada tasks have a ptid_t of the form (PID, 0, TID) where TID
is the Ada task's ID as extracted from Ada runtime information.
Switching to a given Ada task (or its underlying thread) is performed
by fetching the registers of that task from the memory area where
the registers were saved. For any of the other operations, the
operation is performed by first finding the CPU on which the task
is running, switching to its corresponding ptid, and then performing
the operation on that ptid using the target beneath us. */
/* If true, ravenscar task support is enabled. */
static bool ravenscar_task_support = true;
static const char running_thread_name[] = "__gnat_running_thread_table";
static const char known_tasks_name[] = "system__tasking__debug__known_tasks";
static const char first_task_name[] = "system__tasking__debug__first_task";
static const char ravenscar_runtime_initializer[]
= "system__bb__threads__initialize";
static const target_info ravenscar_target_info = {
"ravenscar",
N_("Ravenscar tasks."),
N_("Ravenscar tasks support.")
};
struct ravenscar_thread_target final : public target_ops
{
ravenscar_thread_target ()
: m_base_ptid (inferior_ptid)
{
}
const target_info &info () const override
{ return ravenscar_target_info; }
strata stratum () const override { return thread_stratum; }
ptid_t wait (ptid_t, struct target_waitstatus *, target_wait_flags) override;
void resume (ptid_t, int, enum gdb_signal) override;
void fetch_registers (struct regcache *, int) override;
void store_registers (struct regcache *, int) override;
void prepare_to_store (struct regcache *) override;
bool stopped_by_sw_breakpoint () override;
bool stopped_by_hw_breakpoint () override;
bool stopped_by_watchpoint () override;
bool stopped_data_address (CORE_ADDR *) override;
enum target_xfer_status xfer_partial (enum target_object object,
const char *annex,
gdb_byte *readbuf,
const gdb_byte *writebuf,
ULONGEST offset, ULONGEST len,
ULONGEST *xfered_len) override;
bool thread_alive (ptid_t ptid) override;
int core_of_thread (ptid_t ptid) override;
void update_thread_list () override;
std::string pid_to_str (ptid_t) override;
ptid_t get_ada_task_ptid (long lwp, ULONGEST thread) override;
struct btrace_target_info *enable_btrace (thread_info *tp,
const struct btrace_config *conf)
override
{
process_stratum_target *proc_target
= as_process_stratum_target (this->beneath ());
ptid_t underlying = get_base_thread_from_ravenscar_task (tp->ptid);
tp = proc_target->find_thread (underlying);
return beneath ()->enable_btrace (tp, conf);
}
void mourn_inferior () override;
void close () override
{
delete this;
}
thread_info *add_active_thread ();
private:
/* PTID of the last thread that received an event.
This can be useful to determine the associated task that received
the event, to make it the current task. */
ptid_t m_base_ptid;
ptid_t active_task (int cpu);
bool task_is_currently_active (ptid_t ptid);
bool runtime_initialized ();
int get_thread_base_cpu (ptid_t ptid);
ptid_t get_base_thread_from_ravenscar_task (ptid_t ptid);
void add_thread (struct ada_task_info *task);
/* Like switch_to_thread, but uses the base ptid for the thread. */
void set_base_thread_from_ravenscar_task (ptid_t ptid)
{
process_stratum_target *proc_target
= as_process_stratum_target (this->beneath ());
ptid_t underlying = get_base_thread_from_ravenscar_task (ptid);
switch_to_thread (proc_target->find_thread (underlying));
}
/* Some targets use lazy FPU initialization. On these, the FP
registers for a given task might be uninitialized, or stored in
the per-task context, or simply be the live registers on the CPU.
This enum is used to encode this information. */
enum fpu_state
{
/* This target doesn't do anything special for FP registers -- if
any exist, they are treated just identical to non-FP
registers. */
NOTHING_SPECIAL,
/* This target uses the lazy FP scheme, and the FP registers are
taken from the CPU. This can happen for any task, because if a
task switch occurs, the registers aren't immediately written to
the per-task context -- this is deferred until the current task
causes an FPU trap. */
LIVE_FP_REGISTERS,
/* This target uses the lazy FP scheme, and the FP registers are
not available. Maybe this task never initialized the FPU, or
maybe GDB couldn't find the required symbol. */
NO_FP_REGISTERS
};
/* Return the FPU state. */
fpu_state get_fpu_state (struct regcache *regcache,
const ravenscar_arch_ops *arch_ops);
/* This maps a TID to the CPU on which it was running. This is
needed because sometimes the runtime will report an active task
that hasn't yet been put on the list of tasks that is read by
ada-tasks.c. */
std::unordered_map<ULONGEST, int> m_cpu_map;
};
/* Return true iff PTID corresponds to a ravenscar task. */
static bool
is_ravenscar_task (ptid_t ptid)
{
/* By construction, ravenscar tasks have their LWP set to zero.
Also make sure that the TID is nonzero, as some remotes, when
asked for the list of threads, will return the first thread
as having its TID set to zero. For instance, TSIM version
2.0.48 for LEON3 sends 'm0' as a reply to the 'qfThreadInfo'
query, which the remote protocol layer then treats as a thread
whose TID is 0. This is obviously not a ravenscar task. */
return ptid.lwp () == 0 && ptid.tid () != 0;
}
/* Given PTID, which can be either a ravenscar task or a CPU thread,
return which CPU that ptid is running on.
This assume that PTID is a valid ptid_t. Otherwise, a gdb_assert
will be triggered. */
int
ravenscar_thread_target::get_thread_base_cpu (ptid_t ptid)
{
int base_cpu;
if (is_ravenscar_task (ptid))
{
/* Prefer to not read inferior memory if possible, to avoid
reentrancy problems with xfer_partial. */
auto iter = m_cpu_map.find (ptid.tid ());
if (iter != m_cpu_map.end ())
base_cpu = iter->second;
else
{
struct ada_task_info *task_info = ada_get_task_info_from_ptid (ptid);
gdb_assert (task_info != NULL);
base_cpu = task_info->base_cpu;
}
}
else
{
/* We assume that the LWP of the PTID is equal to the CPU number. */
base_cpu = ptid.lwp ();
}
return base_cpu;
}
/* Given a ravenscar task (identified by its ptid_t PTID), return true
if this task is the currently active task on the cpu that task is
running on.
In other words, this function determine which CPU this task is
currently running on, and then return nonzero if the CPU in question
is executing the code for that task. If that's the case, then
that task's registers are in the CPU bank. Otherwise, the task
is currently suspended, and its registers have been saved in memory. */
bool
ravenscar_thread_target::task_is_currently_active (ptid_t ptid)
{
ptid_t active_task_ptid = active_task (get_thread_base_cpu (ptid));
return ptid == active_task_ptid;
}
/* Return the CPU thread (as a ptid_t) on which the given ravenscar
task is running.
This is the thread that corresponds to the CPU on which the task
is running. */
ptid_t
ravenscar_thread_target::get_base_thread_from_ravenscar_task (ptid_t ptid)
{
int base_cpu;
if (!is_ravenscar_task (ptid))
return ptid;
base_cpu = get_thread_base_cpu (ptid);
return ptid_t (ptid.pid (), base_cpu);
}
/* Fetch the ravenscar running thread from target memory, make sure
there's a corresponding thread in the thread list, and return it.
If the runtime is not initialized, return NULL. */
thread_info *
ravenscar_thread_target::add_active_thread ()
{
process_stratum_target *proc_target
= as_process_stratum_target (this->beneath ());
int base_cpu;
gdb_assert (!is_ravenscar_task (m_base_ptid));
base_cpu = get_thread_base_cpu (m_base_ptid);
if (!runtime_initialized ())
return nullptr;
/* It's possible for runtime_initialized to return true but for it
not to be fully initialized. For example, this can happen for a
breakpoint placed at the task's beginning. */
ptid_t active_ptid = active_task (base_cpu);
if (active_ptid == null_ptid)
return nullptr;
/* The running thread may not have been added to
system.tasking.debug's list yet; so ravenscar_update_thread_list
may not always add it to the thread list. Add it here. */
thread_info *active_thr = proc_target->find_thread (active_ptid);
if (active_thr == nullptr)
{
active_thr = ::add_thread (proc_target, active_ptid);
m_cpu_map[active_ptid.tid ()] = base_cpu;
}
return active_thr;
}
/* The Ravenscar Runtime exports a symbol which contains the ID of
the thread that is currently running. Try to locate that symbol
and return its associated minimal symbol.
Return NULL if not found. */
static struct bound_minimal_symbol
get_running_thread_msymbol ()
{
struct bound_minimal_symbol msym;
msym = lookup_minimal_symbol (running_thread_name, NULL, NULL);
if (!msym.minsym)
/* Older versions of the GNAT runtime were using a different
(less ideal) name for the symbol where the active thread ID
is stored. If we couldn't find the symbol using the latest
name, then try the old one. */
msym = lookup_minimal_symbol ("running_thread", NULL, NULL);
return msym;
}
/* Return True if the Ada Ravenscar run-time can be found in the
application. */
static bool
has_ravenscar_runtime ()
{
struct bound_minimal_symbol msym_ravenscar_runtime_initializer
= lookup_minimal_symbol (ravenscar_runtime_initializer, NULL, NULL);
struct bound_minimal_symbol msym_known_tasks
= lookup_minimal_symbol (known_tasks_name, NULL, NULL);
struct bound_minimal_symbol msym_first_task
= lookup_minimal_symbol (first_task_name, NULL, NULL);
struct bound_minimal_symbol msym_running_thread
= get_running_thread_msymbol ();
return (msym_ravenscar_runtime_initializer.minsym
&& (msym_known_tasks.minsym || msym_first_task.minsym)
&& msym_running_thread.minsym);
}
/* Return True if the Ada Ravenscar run-time can be found in the
application, and if it has been initialized on target. */
bool
ravenscar_thread_target::runtime_initialized ()
{
return active_task (1) != null_ptid;
}
/* Return the ID of the thread that is currently running.
Return 0 if the ID could not be determined. */
static CORE_ADDR
get_running_thread_id (int cpu)
{
struct bound_minimal_symbol object_msym = get_running_thread_msymbol ();
int object_size;
int buf_size;
gdb_byte *buf;
CORE_ADDR object_addr;
struct type *builtin_type_void_data_ptr
= builtin_type (current_inferior ()->arch ())->builtin_data_ptr;
if (!object_msym.minsym)
return 0;
object_size = builtin_type_void_data_ptr->length ();
object_addr = (object_msym.value_address ()
+ (cpu - 1) * object_size);
buf_size = object_size;
buf = (gdb_byte *) alloca (buf_size);
read_memory (object_addr, buf, buf_size);
return extract_typed_address (buf, builtin_type_void_data_ptr);
}
void
ravenscar_thread_target::resume (ptid_t ptid, int step,
enum gdb_signal siggnal)
{
/* If we see a wildcard resume, we simply pass that on. Otherwise,
arrange to resume the base ptid. */
inferior_ptid = m_base_ptid;
if (ptid.is_pid ())
{
/* We only have one process, so resume all threads of it. */
ptid = minus_one_ptid;
}
else if (ptid != minus_one_ptid)
ptid = m_base_ptid;
beneath ()->resume (ptid, step, siggnal);
}
ptid_t
ravenscar_thread_target::wait (ptid_t ptid,
struct target_waitstatus *status,
target_wait_flags options)
{
process_stratum_target *beneath
= as_process_stratum_target (this->beneath ());
ptid_t event_ptid;
if (ptid != minus_one_ptid)
ptid = m_base_ptid;
event_ptid = beneath->wait (ptid, status, 0);
/* Find any new threads that might have been created, and return the
active thread.
Only do it if the program is still alive, though. Otherwise,
this causes problems when debugging through the remote protocol,
because we might try switching threads (and thus sending packets)
after the remote has disconnected. */
if (status->kind () != TARGET_WAITKIND_EXITED
&& status->kind () != TARGET_WAITKIND_SIGNALLED
&& runtime_initialized ())
{
m_base_ptid = event_ptid;
this->update_thread_list ();
thread_info *thr = this->add_active_thread ();
if (thr != nullptr)
return thr->ptid;
}
return event_ptid;
}
/* Add the thread associated to the given TASK to the thread list
(if the thread has already been added, this is a no-op). */
void
ravenscar_thread_target::add_thread (struct ada_task_info *task)
{
if (current_inferior ()->find_thread (task->ptid) == NULL)
{
::add_thread (current_inferior ()->process_target (), task->ptid);
m_cpu_map[task->ptid.tid ()] = task->base_cpu;
}
}
void
ravenscar_thread_target::update_thread_list ()
{
/* iterate_over_live_ada_tasks requires that inferior_ptid be set,
but this isn't always the case in target methods. So, we ensure
it here. */
scoped_restore save_ptid = make_scoped_restore (&inferior_ptid,
m_base_ptid);
/* Do not clear the thread list before adding the Ada task, to keep
the thread that the process stratum has included into it
(m_base_ptid) and the running thread, that may not have been included
to system.tasking.debug's list yet. */
iterate_over_live_ada_tasks ([this] (struct ada_task_info *task)
{
this->add_thread (task);
});
}
ptid_t
ravenscar_thread_target::active_task (int cpu)
{
CORE_ADDR tid = get_running_thread_id (cpu);
if (tid == 0)
return null_ptid;
else
return ptid_t (m_base_ptid.pid (), 0, tid);
}
bool
ravenscar_thread_target::thread_alive (ptid_t ptid)
{
/* Ravenscar tasks are non-terminating. */
return true;
}
std::string
ravenscar_thread_target::pid_to_str (ptid_t ptid)
{
if (!is_ravenscar_task (ptid))
return beneath ()->pid_to_str (ptid);
return string_printf ("Ravenscar Thread 0x%s",
phex_nz (ptid.tid (), sizeof (ULONGEST)));
}
CORE_ADDR
ravenscar_arch_ops::get_stack_base (struct regcache *regcache) const
{
struct gdbarch *gdbarch = regcache->arch ();
const int sp_regnum = gdbarch_sp_regnum (gdbarch);
ULONGEST stack_address;
regcache_cooked_read_unsigned (regcache, sp_regnum, &stack_address);
return (CORE_ADDR) stack_address;
}
void
ravenscar_arch_ops::supply_one_register (struct regcache *regcache,
int regnum,
CORE_ADDR descriptor,
CORE_ADDR stack_base) const
{
CORE_ADDR addr;
if (regnum >= first_stack_register && regnum <= last_stack_register)
addr = stack_base;
else
addr = descriptor;
addr += offsets[regnum];
struct gdbarch *gdbarch = regcache->arch ();
int size = register_size (gdbarch, regnum);
gdb_byte *buf = (gdb_byte *) alloca (size);
read_memory (addr, buf, size);
regcache->raw_supply (regnum, buf);
}
void
ravenscar_arch_ops::fetch_register (struct regcache *regcache,
int regnum) const
{
gdb_assert (regnum != -1);
struct gdbarch *gdbarch = regcache->arch ();
/* The tid is the thread_id field, which is a pointer to the thread. */
CORE_ADDR thread_descriptor_address
= (CORE_ADDR) regcache->ptid ().tid ();
int sp_regno = -1;
CORE_ADDR stack_address = 0;
if (regnum >= first_stack_register && regnum <= last_stack_register)
{
/* We must supply SP for get_stack_base, so recurse. */
sp_regno = gdbarch_sp_regnum (gdbarch);
gdb_assert (!(sp_regno >= first_stack_register
&& sp_regno <= last_stack_register));
fetch_register (regcache, sp_regno);
stack_address = get_stack_base (regcache);
}
if (regnum < offsets.size () && offsets[regnum] != -1)
supply_one_register (regcache, regnum, thread_descriptor_address,
stack_address);
}
void
ravenscar_arch_ops::store_one_register (struct regcache *regcache, int regnum,
CORE_ADDR descriptor,
CORE_ADDR stack_base) const
{
CORE_ADDR addr;
if (regnum >= first_stack_register && regnum <= last_stack_register)
addr = stack_base;
else
addr = descriptor;
addr += offsets[regnum];
struct gdbarch *gdbarch = regcache->arch ();
int size = register_size (gdbarch, regnum);
gdb_byte *buf = (gdb_byte *) alloca (size);
regcache->raw_collect (regnum, buf);
write_memory (addr, buf, size);
}
void
ravenscar_arch_ops::store_register (struct regcache *regcache,
int regnum) const
{
gdb_assert (regnum != -1);
/* The tid is the thread_id field, which is a pointer to the thread. */
CORE_ADDR thread_descriptor_address
= (CORE_ADDR) regcache->ptid ().tid ();
CORE_ADDR stack_address = 0;
if (regnum >= first_stack_register && regnum <= last_stack_register)
stack_address = get_stack_base (regcache);
if (regnum < offsets.size () && offsets[regnum] != -1)
store_one_register (regcache, regnum, thread_descriptor_address,
stack_address);
}
/* Temporarily set the ptid of a regcache to some other value. When
this object is destroyed, the regcache's original ptid is
restored. */
class temporarily_change_regcache_ptid
{
public:
temporarily_change_regcache_ptid (struct regcache *regcache, ptid_t new_ptid)
: m_regcache (regcache),
m_save_ptid (regcache->ptid ())
{
m_regcache->set_ptid (new_ptid);
}
~temporarily_change_regcache_ptid ()
{
m_regcache->set_ptid (m_save_ptid);
}
private:
/* The regcache. */
struct regcache *m_regcache;
/* The saved ptid. */
ptid_t m_save_ptid;
};
ravenscar_thread_target::fpu_state
ravenscar_thread_target::get_fpu_state (struct regcache *regcache,
const ravenscar_arch_ops *arch_ops)
{
/* We want to return true if the special FP register handling is
needed. If this target doesn't have lazy FP, then no special
treatment is ever needed. */
if (!arch_ops->on_demand_fp ())
return NOTHING_SPECIAL;
bound_minimal_symbol fpu_context
= lookup_minimal_symbol ("system__bb__cpu_primitives__current_fpu_context",
nullptr, nullptr);
/* If the symbol can't be found, just fall back. */
if (fpu_context.minsym == nullptr)
return NO_FP_REGISTERS;
type *ptr_type
= builtin_type (current_inferior ()->arch ())->builtin_data_ptr;
ptr_type = lookup_pointer_type (ptr_type);
value *val = value_from_pointer (ptr_type, fpu_context.value_address ());
int cpu = get_thread_base_cpu (regcache->ptid ());
/* The array index type has a lower bound of 1 -- it is Ada code --
so subtract 1 here. */
val = value_ptradd (val, cpu - 1);
val = value_ind (val);
CORE_ADDR fpu_task = value_as_long (val);
/* The tid is the thread_id field, which is a pointer to the thread. */
CORE_ADDR thread_descriptor_address
= (CORE_ADDR) regcache->ptid ().tid ();
if (fpu_task == (thread_descriptor_address
+ arch_ops->get_fpu_context_offset ()))
return LIVE_FP_REGISTERS;
int v_init_offset = arch_ops->get_v_init_offset ();
gdb_byte init = 0;
read_memory (thread_descriptor_address + v_init_offset, &init, 1);
return init ? NOTHING_SPECIAL : NO_FP_REGISTERS;
}
void
ravenscar_thread_target::fetch_registers (struct regcache *regcache,
int regnum)
{
ptid_t ptid = regcache->ptid ();
if (runtime_initialized () && is_ravenscar_task (ptid))
{
struct gdbarch *gdbarch = regcache->arch ();
bool is_active = task_is_currently_active (ptid);
struct ravenscar_arch_ops *arch_ops = gdbarch_ravenscar_ops (gdbarch);
gdb::optional<fpu_state> fp_state;
int low_reg = regnum == -1 ? 0 : regnum;
int high_reg = regnum == -1 ? gdbarch_num_regs (gdbarch) : regnum + 1;
ptid_t base = get_base_thread_from_ravenscar_task (ptid);
for (int i = low_reg; i < high_reg; ++i)
{
bool use_beneath = false;
if (arch_ops->is_fp_register (i))
{
if (!fp_state.has_value ())
fp_state = get_fpu_state (regcache, arch_ops);
if (*fp_state == NO_FP_REGISTERS)
continue;
if (*fp_state == LIVE_FP_REGISTERS
|| (is_active && *fp_state == NOTHING_SPECIAL))
use_beneath = true;
}
else
use_beneath = is_active;
if (use_beneath)
{
temporarily_change_regcache_ptid changer (regcache, base);
beneath ()->fetch_registers (regcache, i);
}
else
arch_ops->fetch_register (regcache, i);
}
}
else
beneath ()->fetch_registers (regcache, regnum);
}
void
ravenscar_thread_target::store_registers (struct regcache *regcache,
int regnum)
{
ptid_t ptid = regcache->ptid ();
if (runtime_initialized () && is_ravenscar_task (ptid))
{
struct gdbarch *gdbarch = regcache->arch ();
bool is_active = task_is_currently_active (ptid);
struct ravenscar_arch_ops *arch_ops = gdbarch_ravenscar_ops (gdbarch);
gdb::optional<fpu_state> fp_state;
int low_reg = regnum == -1 ? 0 : regnum;
int high_reg = regnum == -1 ? gdbarch_num_regs (gdbarch) : regnum + 1;
ptid_t base = get_base_thread_from_ravenscar_task (ptid);
for (int i = low_reg; i < high_reg; ++i)
{
bool use_beneath = false;
if (arch_ops->is_fp_register (i))
{
if (!fp_state.has_value ())
fp_state = get_fpu_state (regcache, arch_ops);
if (*fp_state == NO_FP_REGISTERS)
continue;
if (*fp_state == LIVE_FP_REGISTERS
|| (is_active && *fp_state == NOTHING_SPECIAL))
use_beneath = true;
}
else
use_beneath = is_active;
if (use_beneath)
{
temporarily_change_regcache_ptid changer (regcache, base);
beneath ()->store_registers (regcache, i);
}
else
arch_ops->store_register (regcache, i);
}
}
else
beneath ()->store_registers (regcache, regnum);
}
void
ravenscar_thread_target::prepare_to_store (struct regcache *regcache)
{
ptid_t ptid = regcache->ptid ();
if (runtime_initialized () && is_ravenscar_task (ptid))
{
if (task_is_currently_active (ptid))
{
ptid_t base = get_base_thread_from_ravenscar_task (ptid);
temporarily_change_regcache_ptid changer (regcache, base);
beneath ()->prepare_to_store (regcache);
}
else
{
/* Nothing. */
}
}
else
beneath ()->prepare_to_store (regcache);
}
/* Implement the to_stopped_by_sw_breakpoint target_ops "method". */
bool
ravenscar_thread_target::stopped_by_sw_breakpoint ()
{
scoped_restore_current_thread saver;
set_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->stopped_by_sw_breakpoint ();
}
/* Implement the to_stopped_by_hw_breakpoint target_ops "method". */
bool
ravenscar_thread_target::stopped_by_hw_breakpoint ()
{
scoped_restore_current_thread saver;
set_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->stopped_by_hw_breakpoint ();
}
/* Implement the to_stopped_by_watchpoint target_ops "method". */
bool
ravenscar_thread_target::stopped_by_watchpoint ()
{
scoped_restore_current_thread saver;
set_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->stopped_by_watchpoint ();
}
/* Implement the to_stopped_data_address target_ops "method". */
bool
ravenscar_thread_target::stopped_data_address (CORE_ADDR *addr_p)
{
scoped_restore_current_thread saver;
set_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->stopped_data_address (addr_p);
}
void
ravenscar_thread_target::mourn_inferior ()
{
m_base_ptid = null_ptid;
target_ops *beneath = this->beneath ();
current_inferior ()->unpush_target (this);
beneath->mourn_inferior ();
}
/* Implement the to_core_of_thread target_ops "method". */
int
ravenscar_thread_target::core_of_thread (ptid_t ptid)
{
scoped_restore_current_thread saver;
set_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->core_of_thread (inferior_ptid);
}
/* Implement the target xfer_partial method. */
enum target_xfer_status
ravenscar_thread_target::xfer_partial (enum target_object object,
const char *annex,
gdb_byte *readbuf,
const gdb_byte *writebuf,
ULONGEST offset, ULONGEST len,
ULONGEST *xfered_len)
{
scoped_restore save_ptid = make_scoped_restore (&inferior_ptid);
/* Calling get_base_thread_from_ravenscar_task can read memory from
the inferior. However, that function is written to prefer our
internal map, so it should not result in recursive calls in
practice. */
inferior_ptid = get_base_thread_from_ravenscar_task (inferior_ptid);
return beneath ()->xfer_partial (object, annex, readbuf, writebuf,
offset, len, xfered_len);
}
/* Observer on inferior_created: push ravenscar thread stratum if needed. */
static void
ravenscar_inferior_created (inferior *inf)
{
const char *err_msg;
if (!ravenscar_task_support
|| gdbarch_ravenscar_ops (current_inferior ()->arch ()) == NULL
|| !has_ravenscar_runtime ())
return;
err_msg = ada_get_tcb_types_info ();
if (err_msg != NULL)
{
warning (_("%s. Task/thread support disabled."), err_msg);
return;
}
ravenscar_thread_target *rtarget = new ravenscar_thread_target ();
inf->push_target (target_ops_up (rtarget));
thread_info *thr = rtarget->add_active_thread ();
if (thr != nullptr)
switch_to_thread (thr);
}
ptid_t
ravenscar_thread_target::get_ada_task_ptid (long lwp, ULONGEST thread)
{
return ptid_t (m_base_ptid.pid (), 0, thread);
}
/* Command-list for the "set/show ravenscar" prefix command. */
static struct cmd_list_element *set_ravenscar_list;
static struct cmd_list_element *show_ravenscar_list;
/* Implement the "show ravenscar task-switching" command. */
static void
show_ravenscar_task_switching_command (struct ui_file *file, int from_tty,
struct cmd_list_element *c,
const char *value)
{
if (ravenscar_task_support)
gdb_printf (file, _("\
Support for Ravenscar task/thread switching is enabled\n"));
else
gdb_printf (file, _("\
Support for Ravenscar task/thread switching is disabled\n"));
}
/* Module startup initialization function, automagically called by
init.c. */
void _initialize_ravenscar ();
void
_initialize_ravenscar ()
{
/* Notice when the inferior is created in order to push the
ravenscar ops if needed. */
gdb::observers::inferior_created.attach (ravenscar_inferior_created,
"ravenscar-thread");
add_setshow_prefix_cmd
("ravenscar", no_class,
_("Prefix command for changing Ravenscar-specific settings."),
_("Prefix command for showing Ravenscar-specific settings."),
&set_ravenscar_list, &show_ravenscar_list,
&setlist, &showlist);
add_setshow_boolean_cmd ("task-switching", class_obscure,
&ravenscar_task_support, _("\
Enable or disable support for GNAT Ravenscar tasks."), _("\
Show whether support for GNAT Ravenscar tasks is enabled."),
_("\
Enable or disable support for task/thread switching with the GNAT\n\
Ravenscar run-time library for bareboard configuration."),
NULL, show_ravenscar_task_switching_command,
&set_ravenscar_list, &show_ravenscar_list);
}
|