1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
|
/* Classes for modeling the state of memory.
Copyright (C) 2020-2024 Free Software Foundation, Inc.
Contributed by David Malcolm <dmalcolm@redhat.com>.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3, or (at your option)
any later version.
GCC is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
General Public License for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
#ifndef GCC_ANALYZER_STORE_H
#define GCC_ANALYZER_STORE_H
/* Implementation of the region-based ternary model described in:
"A Memory Model for Static Analysis of C Programs"
(Zhongxing Xu, Ted Kremenek, and Jian Zhang)
http://lcs.ios.ac.cn/~xuzb/canalyze/memmodel.pdf */
/* The store models memory as a collection of "clusters", where regions
are partitioned into clusters via their base region.
For example, given:
int a, b, c;
struct coord { double x; double y; } verts[3];
then "verts[0].y" and "verts[1].x" both have "verts" as their base region.
Each of a, b, c, and verts will have their own clusters, so that we
know that writes to e.g. "verts[1].x".don't affect e.g. "a".
Within each cluster we store a map of bindings to values, where the
binding keys can be either concrete or symbolic.
Concrete bindings affect a specific range of bits relative to the start
of the base region of the cluster, whereas symbolic bindings affect
a specific subregion within the cluster.
Consider (from the symbolic-1.c testcase):
char arr[1024];
arr[2] = a; (1)
arr[3] = b; (2)
After (1) and (2), the cluster for "arr" has concrete bindings
for bits 16-23 and for bits 24-31, with svalues "INIT_VAL(a)"
and "INIT_VAL(b)" respectively:
cluster: {bits 16-23: "INIT_VAL(a)",
bits 24-31: "INIT_VAL(b)";
flags: {}}
Attempting to query unbound subregions e.g. arr[4] will
return "UNINITIALIZED".
"a" and "b" are each in their own clusters, with no explicit
bindings, and thus implicitly have value INIT_VAL(a) and INIT_VAL(b).
arr[3] = c; (3)
After (3), the concrete binding for bits 24-31 is replaced with the
svalue "INIT_VAL(c)":
cluster: {bits 16-23: "INIT_VAL(a)", (from before)
bits 24-31: "INIT_VAL(c)"; (updated)
flags: {}}
arr[i] = d; (4)
After (4), we lose the concrete bindings and replace them with a
symbolic binding for "arr[i]", with svalue "INIT_VAL(d)". We also
mark the cluster as having been "symbolically touched": future
attempts to query the values of subregions other than "arr[i]",
such as "arr[3]" are "UNKNOWN", since we don't know if the write
to arr[i] affected them.
cluster: {symbolic_key(arr[i]): "INIT_VAL(d)";
flags: {TOUCHED}}
arr[j] = e; (5)
After (5), we lose the symbolic binding for "arr[i]" since we could
have overwritten it, and add a symbolic binding for "arr[j]".
cluster: {symbolic_key(arr[j]): "INIT_VAL(d)"; (different symbolic
flags: {TOUCHED}} binding)
arr[3] = f; (6)
After (6), we lose the symbolic binding for "arr[j]" since we could
have overwritten it, and gain a concrete binding for bits 24-31
again, this time with svalue "INIT_VAL(e)":
cluster: {bits 24-31: "INIT_VAL(d)";
flags: {TOUCHED}}
The cluster is still flagged as touched, so that we know that
accesses to other elements are "UNKNOWN" rather than
"UNINITIALIZED".
Handling symbolic regions requires us to handle aliasing.
In the first example above, each of a, b, c and verts are non-symbolic
base regions and so their clusters are "concrete clusters", whereas given:
struct coord *p, *q;
then "*p" and "*q" are symbolic base regions, and thus "*p" and "*q"
have "symbolic clusters".
In the above, "verts[i].x" will have a symbolic *binding* within a
concrete cluster for "verts", whereas "*p" is a symbolic *cluster*.
Writes to concrete clusters can't affect other concrete clusters,
but can affect symbolic clusters; e.g. after:
verts[0].x = 42;
we bind 42 in the cluster for "verts", but the clusters for "b" and "c"
can't be affected. Any symbolic clusters for *p and for *q can be
affected, *p and *q could alias verts.
Writes to a symbolic cluster can affect other clusters, both
concrete and symbolic; e.g. after:
p->x = 17;
we bind 17 within the cluster for "*p". The concrete clusters for a, b,
c, and verts could be affected, depending on whether *p aliases them.
Similarly, the symbolic cluster to *q could be affected. */
namespace ana {
/* A class for keeping track of aspects of a program_state that we don't
know about, to avoid false positives about leaks.
Consider:
p->field = malloc (1024);
q->field = NULL;
where we don't know whether or not p and q point to the same memory,
and:
p->field = malloc (1024);
unknown_fn (p);
In both cases, the svalue for the address of the allocated buffer
goes from being bound to p->field to not having anything explicitly bound
to it.
Given that we conservatively discard bindings due to possible aliasing or
calls to unknown function, the store loses references to svalues,
but these svalues could still be live. We don't want to warn about
them leaking - they're effectively in a "maybe live" state.
This "maybe live" information is somewhat transient.
We don't want to store this "maybe live" information in the program_state,
region_model, or store, since we don't want to bloat these objects (and
potentially bloat the exploded_graph with more nodes).
However, we can't store it in the region_model_context, as these context
objects sometimes don't last long enough to be around when comparing the
old vs the new state.
This class is a way to track a set of such svalues, so that we can
temporarily capture that they are in a "maybe live" state whilst
comparing old and new states. */
class uncertainty_t
{
public:
typedef hash_set<const svalue *>::iterator iterator;
void on_maybe_bound_sval (const svalue *sval)
{
m_maybe_bound_svals.add (sval);
}
void on_mutable_sval_at_unknown_call (const svalue *sval)
{
m_mutable_at_unknown_call_svals.add (sval);
}
bool unknown_sm_state_p (const svalue *sval)
{
return (m_maybe_bound_svals.contains (sval)
|| m_mutable_at_unknown_call_svals.contains (sval));
}
void dump_to_pp (pretty_printer *pp, bool simple) const;
void dump (bool simple) const;
iterator begin_maybe_bound_svals () const
{
return m_maybe_bound_svals.begin ();
}
iterator end_maybe_bound_svals () const
{
return m_maybe_bound_svals.end ();
}
private:
/* svalues that might or might not still be bound. */
hash_set<const svalue *> m_maybe_bound_svals;
/* svalues that have mutable sm-state at unknown calls. */
hash_set<const svalue *> m_mutable_at_unknown_call_svals;
};
class byte_range;
class concrete_binding;
class symbolic_binding;
/* Abstract base class for describing ranges of bits within a binding_map
that can have svalues bound to them. */
class binding_key
{
public:
virtual ~binding_key () {}
virtual bool concrete_p () const = 0;
bool symbolic_p () const { return !concrete_p (); }
static const binding_key *make (store_manager *mgr, const region *r);
virtual void dump_to_pp (pretty_printer *pp, bool simple) const = 0;
void dump (bool simple) const;
label_text get_desc (bool simple=true) const;
static int cmp_ptrs (const void *, const void *);
static int cmp (const binding_key *, const binding_key *);
virtual const concrete_binding *dyn_cast_concrete_binding () const
{ return NULL; }
virtual const symbolic_binding *dyn_cast_symbolic_binding () const
{ return NULL; }
};
/* A concrete range of bits. */
struct bit_range
{
bit_range (bit_offset_t start_bit_offset, bit_size_t size_in_bits)
: m_start_bit_offset (start_bit_offset),
m_size_in_bits (size_in_bits)
{}
void dump_to_pp (pretty_printer *pp) const;
void dump () const;
json::object *to_json () const;
bool empty_p () const
{
return m_size_in_bits == 0;
}
bit_offset_t get_start_bit_offset () const
{
return m_start_bit_offset;
}
bit_offset_t get_next_bit_offset () const
{
return m_start_bit_offset + m_size_in_bits;
}
bit_offset_t get_last_bit_offset () const
{
gcc_assert (!empty_p ());
return get_next_bit_offset () - 1;
}
bool contains_p (bit_offset_t offset) const
{
return (offset >= get_start_bit_offset ()
&& offset < get_next_bit_offset ());
}
bool contains_p (const bit_range &other, bit_range *out) const;
bool operator== (const bit_range &other) const
{
return (m_start_bit_offset == other.m_start_bit_offset
&& m_size_in_bits == other.m_size_in_bits);
}
bool intersects_p (const bit_range &other) const
{
return (get_start_bit_offset () < other.get_next_bit_offset ()
&& other.get_start_bit_offset () < get_next_bit_offset ());
}
bool intersects_p (const bit_range &other,
bit_size_t *out_num_overlap_bits) const;
bool intersects_p (const bit_range &other,
bit_range *out_this,
bit_range *out_other) const;
bool exceeds_p (const bit_range &other,
bit_range *out_overhanging_bit_range) const;
bool falls_short_of_p (bit_offset_t offset,
bit_range *out_fall_short_bits) const;
static int cmp (const bit_range &br1, const bit_range &br2);
bit_range operator- (bit_offset_t offset) const;
static bool from_mask (unsigned HOST_WIDE_INT mask, bit_range *out);
bool as_byte_range (byte_range *out) const;
bit_offset_t m_start_bit_offset;
bit_size_t m_size_in_bits;
};
/* A concrete range of bytes. */
struct byte_range
{
byte_range (byte_offset_t start_byte_offset, byte_size_t size_in_bytes)
: m_start_byte_offset (start_byte_offset),
m_size_in_bytes (size_in_bytes)
{}
void dump_to_pp (pretty_printer *pp) const;
void dump () const;
json::object *to_json () const;
bool empty_p () const
{
return m_size_in_bytes == 0;
}
bool contains_p (byte_offset_t offset) const
{
return (offset >= get_start_byte_offset ()
&& offset < get_next_byte_offset ());
}
bool contains_p (const byte_range &other, byte_range *out) const;
bool operator== (const byte_range &other) const
{
return (m_start_byte_offset == other.m_start_byte_offset
&& m_size_in_bytes == other.m_size_in_bytes);
}
byte_offset_t get_start_byte_offset () const
{
return m_start_byte_offset;
}
byte_offset_t get_next_byte_offset () const
{
return m_start_byte_offset + m_size_in_bytes;
}
byte_offset_t get_last_byte_offset () const
{
gcc_assert (!empty_p ());
return m_start_byte_offset + m_size_in_bytes - 1;
}
bit_range as_bit_range () const
{
return bit_range (m_start_byte_offset * BITS_PER_UNIT,
m_size_in_bytes * BITS_PER_UNIT);
}
bit_offset_t get_start_bit_offset () const
{
return m_start_byte_offset * BITS_PER_UNIT;
}
bit_offset_t get_next_bit_offset () const
{
return get_next_byte_offset () * BITS_PER_UNIT;
}
static int cmp (const byte_range &br1, const byte_range &br2);
byte_offset_t m_start_byte_offset;
byte_size_t m_size_in_bytes;
};
/* Concrete subclass of binding_key, for describing a non-empty
concrete range of bits within the binding_map (e.g. "bits 8-15"). */
class concrete_binding : public binding_key
{
public:
/* This class is its own key for the purposes of consolidation. */
typedef concrete_binding key_t;
concrete_binding (bit_offset_t start_bit_offset, bit_size_t size_in_bits)
: m_bit_range (start_bit_offset, size_in_bits)
{
gcc_assert (m_bit_range.m_size_in_bits > 0);
}
bool concrete_p () const final override { return true; }
hashval_t hash () const
{
inchash::hash hstate;
hstate.add_wide_int (m_bit_range.m_start_bit_offset);
hstate.add_wide_int (m_bit_range.m_size_in_bits);
return hstate.end ();
}
bool operator== (const concrete_binding &other) const
{
return m_bit_range == other.m_bit_range;
}
void dump_to_pp (pretty_printer *pp, bool simple) const final override;
const concrete_binding *dyn_cast_concrete_binding () const final override
{ return this; }
const bit_range &get_bit_range () const { return m_bit_range; }
bool get_byte_range (byte_range *out) const;
bit_offset_t get_start_bit_offset () const
{
return m_bit_range.m_start_bit_offset;
}
bit_size_t get_size_in_bits () const
{
return m_bit_range.m_size_in_bits;
}
/* Return the next bit offset after the end of this binding. */
bit_offset_t get_next_bit_offset () const
{
return m_bit_range.get_next_bit_offset ();
}
bool overlaps_p (const concrete_binding &other) const;
static int cmp_ptr_ptr (const void *, const void *);
void mark_deleted () { m_bit_range.m_size_in_bits = -1; }
void mark_empty () { m_bit_range.m_size_in_bits = -2; }
bool is_deleted () const { return m_bit_range.m_size_in_bits == -1; }
bool is_empty () const { return m_bit_range.m_size_in_bits == -2; }
private:
bit_range m_bit_range;
};
} // namespace ana
template <>
template <>
inline bool
is_a_helper <const ana::concrete_binding *>::test (const ana::binding_key *key)
{
return key->concrete_p ();
}
template <> struct default_hash_traits<ana::concrete_binding>
: public member_function_hash_traits<ana::concrete_binding>
{
static const bool empty_zero_p = false;
};
namespace ana {
/* Concrete subclass of binding_key, for describing a symbolic set of
bits within the binding_map in terms of a region (e.g. "arr[i]"). */
class symbolic_binding : public binding_key
{
public:
/* This class is its own key for the purposes of consolidation. */
typedef symbolic_binding key_t;
symbolic_binding (const region *region) : m_region (region) {}
bool concrete_p () const final override { return false; }
hashval_t hash () const
{
return (intptr_t)m_region;
}
bool operator== (const symbolic_binding &other) const
{
return m_region == other.m_region;
}
void dump_to_pp (pretty_printer *pp, bool simple) const final override;
const symbolic_binding *dyn_cast_symbolic_binding () const final override
{ return this; }
const region *get_region () const { return m_region; }
static int cmp_ptr_ptr (const void *, const void *);
void mark_deleted () { m_region = reinterpret_cast<const region *> (1); }
void mark_empty () { m_region = NULL; }
bool is_deleted () const
{ return m_region == reinterpret_cast<const region *> (1); }
bool is_empty () const { return m_region == NULL; }
private:
const region *m_region;
};
} // namespace ana
template <> struct default_hash_traits<ana::symbolic_binding>
: public member_function_hash_traits<ana::symbolic_binding>
{
static const bool empty_zero_p = true;
};
namespace ana {
/* A mapping from binding_keys to svalues, for use by binding_cluster
and compound_svalue. */
class binding_map
{
public:
typedef hash_map <const binding_key *, const svalue *> map_t;
typedef map_t::iterator iterator_t;
binding_map () : m_map () {}
binding_map (const binding_map &other);
binding_map& operator=(const binding_map &other);
bool operator== (const binding_map &other) const;
bool operator!= (const binding_map &other) const
{
return !(*this == other);
}
hashval_t hash () const;
const svalue *get (const binding_key *key) const
{
const svalue **slot = const_cast<map_t &> (m_map).get (key);
if (slot)
return *slot;
else
return NULL;
}
bool put (const binding_key *k, const svalue *v)
{
gcc_assert (v);
return m_map.put (k, v);
}
void remove (const binding_key *k) { m_map.remove (k); }
void empty () { m_map.empty (); }
iterator_t begin () const { return m_map.begin (); }
iterator_t end () const { return m_map.end (); }
size_t elements () const { return m_map.elements (); }
void dump_to_pp (pretty_printer *pp, bool simple, bool multiline) const;
void dump (bool simple) const;
json::object *to_json () const;
bool apply_ctor_to_region (const region *parent_reg, tree ctor,
region_model_manager *mgr);
static int cmp (const binding_map &map1, const binding_map &map2);
void remove_overlapping_bindings (store_manager *mgr,
const binding_key *drop_key,
uncertainty_t *uncertainty,
svalue_set *maybe_live_values,
bool always_overlap);
private:
void get_overlapping_bindings (const binding_key *key,
auto_vec<const binding_key *> *out);
bool apply_ctor_val_to_range (const region *parent_reg,
region_model_manager *mgr,
tree min_index, tree max_index,
tree val);
bool apply_ctor_pair_to_child_region (const region *parent_reg,
region_model_manager *mgr,
tree index, tree val);
map_t m_map;
};
/* Concept: BindingVisitor, for use by binding_cluster::for_each_binding
and store::for_each_binding.
Should implement:
void on_binding (const binding_key *key, const svalue *&sval);
*/
/* All of the bindings within a store for regions that share the same
base region. */
class binding_cluster
{
public:
friend class store;
typedef hash_map <const binding_key *, const svalue *> map_t;
typedef map_t::iterator iterator_t;
binding_cluster (const region *base_region);
binding_cluster (const binding_cluster &other);
binding_cluster& operator=(const binding_cluster &other);
bool operator== (const binding_cluster &other) const;
bool operator!= (const binding_cluster &other) const
{
return !(*this == other);
}
hashval_t hash () const;
bool symbolic_p () const;
const region *get_base_region () const { return m_base_region; }
void dump_to_pp (pretty_printer *pp, bool simple, bool multiline) const;
void dump (bool simple) const;
void validate () const;
json::object *to_json () const;
void bind (store_manager *mgr, const region *, const svalue *);
void clobber_region (store_manager *mgr, const region *reg);
void purge_region (store_manager *mgr, const region *reg);
void fill_region (store_manager *mgr, const region *reg, const svalue *sval);
void zero_fill_region (store_manager *mgr, const region *reg);
void mark_region_as_unknown (store_manager *mgr,
const region *reg_to_bind,
const region *reg_for_overlap,
uncertainty_t *uncertainty,
svalue_set *maybe_live_values);
void purge_state_involving (const svalue *sval,
region_model_manager *sval_mgr);
const svalue *get_binding (store_manager *mgr, const region *reg) const;
const svalue *get_binding_recursive (store_manager *mgr,
const region *reg) const;
const svalue *get_any_binding (store_manager *mgr,
const region *reg) const;
const svalue *maybe_get_compound_binding (store_manager *mgr,
const region *reg) const;
void remove_overlapping_bindings (store_manager *mgr, const region *reg,
uncertainty_t *uncertainty,
svalue_set *maybe_live_values);
template <typename T>
void for_each_value (void (*cb) (const svalue *sval, T user_data),
T user_data) const
{
for (map_t::iterator iter = m_map.begin (); iter != m_map.end (); ++iter)
cb ((*iter).second, user_data);
}
static bool can_merge_p (const binding_cluster *cluster_a,
const binding_cluster *cluster_b,
binding_cluster *out_cluster,
store *out_store,
store_manager *mgr,
model_merger *merger);
void make_unknown_relative_to (const binding_cluster *other_cluster,
store *out_store,
store_manager *mgr);
void mark_as_escaped ();
void on_unknown_fncall (const gcall *call, store_manager *mgr,
const conjured_purge &p);
void on_asm (const gasm *stmt, store_manager *mgr,
const conjured_purge &p);
bool escaped_p () const;
bool touched_p () const { return m_touched; }
bool redundant_p () const;
bool empty_p () const { return m_map.elements () == 0; }
void get_representative_path_vars (const region_model *model,
svalue_set *visited,
const region *base_reg,
const svalue *sval,
auto_vec<path_var> *out_pvs) const;
const svalue *maybe_get_simple_value (store_manager *mgr) const;
template <typename BindingVisitor>
void for_each_binding (BindingVisitor &v) const
{
for (map_t::iterator iter = m_map.begin (); iter != m_map.end (); ++iter)
{
const binding_key *key = (*iter).first;
const svalue *&sval = (*iter).second;
v.on_binding (key, sval);
}
}
iterator_t begin () const { return m_map.begin (); }
iterator_t end () const { return m_map.end (); }
const binding_map &get_map () const { return m_map; }
private:
const svalue *get_any_value (const binding_key *key) const;
void bind_compound_sval (store_manager *mgr,
const region *reg,
const compound_svalue *compound_sval);
void bind_key (const binding_key *key, const svalue *sval);
const region *m_base_region;
binding_map m_map;
/* Has a pointer to this cluster "escaped" into a part of the program
we don't know about (via a call to a function with an unknown body,
or by being passed in as a pointer param of a "top-level" function call).
Such regions could be overwritten when other such functions are called,
even if the region is no longer reachable by pointers that we are
tracking. */
bool m_escaped;
/* Has this cluster been written to via a symbolic binding?
If so, then we don't know anything about unbound subregions,
so we can't use initial_svalue, treat them as uninitialized, or
inherit values from a parent region. */
bool m_touched;
};
/* The mapping from regions to svalues.
This is actually expressed by subdividing into clusters, to better
handle aliasing. */
class store
{
public:
typedef hash_map <const region *, binding_cluster *> cluster_map_t;
store ();
store (const store &other);
~store ();
store &operator= (const store &other);
bool operator== (const store &other) const;
bool operator!= (const store &other) const
{
return !(*this == other);
}
hashval_t hash () const;
void dump_to_pp (pretty_printer *pp, bool summarize, bool multiline,
store_manager *mgr) const;
void dump (bool simple) const;
void summarize_to_pp (pretty_printer *pp, bool simple) const;
void validate () const;
json::object *to_json () const;
const svalue *get_any_binding (store_manager *mgr, const region *reg) const;
bool called_unknown_fn_p () const { return m_called_unknown_fn; }
void set_value (store_manager *mgr, const region *lhs_reg,
const svalue *rhs_sval,
uncertainty_t *uncertainty);
void clobber_region (store_manager *mgr, const region *reg);
void purge_region (store_manager *mgr, const region *reg);
void fill_region (store_manager *mgr, const region *reg, const svalue *sval);
void zero_fill_region (store_manager *mgr, const region *reg);
void mark_region_as_unknown (store_manager *mgr, const region *reg,
uncertainty_t *uncertainty,
svalue_set *maybe_live_values);
void purge_state_involving (const svalue *sval,
region_model_manager *sval_mgr);
const binding_cluster *get_cluster (const region *base_reg) const;
binding_cluster *get_cluster (const region *base_reg);
binding_cluster *get_or_create_cluster (const region *base_reg);
void purge_cluster (const region *base_reg);
template <typename T>
void for_each_cluster (void (*cb) (const region *base_reg, T user_data),
T user_data) const
{
for (cluster_map_t::iterator iter = m_cluster_map.begin ();
iter != m_cluster_map.end (); ++iter)
cb ((*iter).first, user_data);
}
static bool can_merge_p (const store *store_a, const store *store_b,
store *out_store, store_manager *mgr,
model_merger *merger);
void mark_as_escaped (const region *base_reg);
void on_unknown_fncall (const gcall *call, store_manager *mgr,
const conjured_purge &p);
bool escaped_p (const region *reg) const;
void get_representative_path_vars (const region_model *model,
svalue_set *visited,
const svalue *sval,
auto_vec<path_var> *out_pvs) const;
cluster_map_t::iterator begin () const { return m_cluster_map.begin (); }
cluster_map_t::iterator end () const { return m_cluster_map.end (); }
tristate eval_alias (const region *base_reg_a,
const region *base_reg_b) const;
template <typename BindingVisitor>
void for_each_binding (BindingVisitor &v)
{
for (cluster_map_t::iterator iter = m_cluster_map.begin ();
iter != m_cluster_map.end (); ++iter)
(*iter).second->for_each_binding (v);
}
void canonicalize (store_manager *mgr);
void loop_replay_fixup (const store *other_store,
region_model_manager *mgr);
void replay_call_summary (call_summary_replay &r,
const store &summary);
void replay_call_summary_cluster (call_summary_replay &r,
const store &summary,
const region *base_reg);
void on_maybe_live_values (const svalue_set &maybe_live_values);
private:
void remove_overlapping_bindings (store_manager *mgr, const region *reg,
uncertainty_t *uncertainty);
tristate eval_alias_1 (const region *base_reg_a,
const region *base_reg_b) const;
cluster_map_t m_cluster_map;
/* If this is true, then unknown code has been called, and so
any global variable that isn't currently modelled by the store
has unknown state, rather than being in an "initial state".
This is to avoid having to mark (and thus explicitly track)
every global when an unknown function is called; instead, they
can be tracked implicitly. */
bool m_called_unknown_fn;
};
/* A class responsible for owning and consolidating binding keys
(both concrete and symbolic).
Key instances are immutable as far as clients are concerned, so they
are provided as "const" ptrs. */
class store_manager
{
public:
store_manager (region_model_manager *mgr) : m_mgr (mgr) {}
logger *get_logger () const;
/* binding consolidation. */
const concrete_binding *
get_concrete_binding (bit_offset_t start_bit_offset,
bit_offset_t size_in_bits);
const concrete_binding *
get_concrete_binding (const bit_range &bits)
{
return get_concrete_binding (bits.get_start_bit_offset (),
bits.m_size_in_bits);
}
const concrete_binding *
get_concrete_binding (const byte_range &bytes)
{
bit_range bits = bytes.as_bit_range ();
return get_concrete_binding (bits);
}
const symbolic_binding *
get_symbolic_binding (const region *region);
region_model_manager *get_svalue_manager () const
{
return m_mgr;
}
void log_stats (logger *logger, bool show_objs) const;
private:
region_model_manager *m_mgr;
consolidation_map<concrete_binding> m_concrete_binding_key_mgr;
consolidation_map<symbolic_binding> m_symbolic_binding_key_mgr;
};
} // namespace ana
#endif /* GCC_ANALYZER_STORE_H */
|