aboutsummaryrefslogtreecommitdiff
path: root/gdb/testsuite/gdb.base/bigcore.exp
blob: 7727e482b824c8df4c58f50e1f3984b2dff195c0 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
# Copyright 1992-2025 Free Software Foundation, Inc.

# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program.  If not, see <http://www.gnu.org/licenses/>.

# This file is based on corefile.exp which was written by Fred
# Fish. (fnf@cygnus.com)


# Are we on a target board?  As of 2004-02-12, GDB didn't have a
# mechanism that would let it efficiently access a remote corefile.

require isnative

# Can the system run this test (in particular support sparse
# corefiles)?  On systems that lack sparse corefile support this test
# consumes too many resources - gigabytes worth of disk space and
# I/O bandwidth.

if { [istarget "*-*-*bsd*"]
     || [istarget "*-*-solaris*"] 
     || [istarget "*-*-darwin*"] 
     || [istarget "*-*-cygwin*"] } {
    untested "kernel lacks sparse corefile support (PR gdb/1551)"
    return
}

standard_testfile .c
set corefile [standard_output_file ${binfile}.corefile]

if  { [gdb_compile "${srcdir}/${subdir}/${srcfile}" "${binfile}" executable {debug}] != "" } {
     untested "failed to compile"
     return -1
}

# Traverse part of bigcore's linked list of memory chunks (forward or
# backward), saving each chunk's address.

proc extract_heap { dir } {
    global gdb_prompt
    global expect_out
    set heap ""
    set test "extract ${dir} heap"
    set lim 0
    gdb_test_multiple "print heap.${dir}" "$test" {
	-re " = \\(struct list \\*\\) 0x0.*$gdb_prompt $" {
	    pass "$test"
	}
	-re " = \\(struct list \\*\\) (0x\[0-9a-f\]*).*$gdb_prompt $" {
	    set heap [concat $heap $expect_out(1,string)]
	    if { $lim >= 200 } {
		pass "$test (stop at $lim)"
	    } else {
		incr lim
		send_gdb "print \$.${dir}\n"
		exp_continue
	    }
	}
	-re ".*$gdb_prompt $" {
	    fail "$test (entry $lim)"
	}
	timeout {
	    fail "$test (timeout)"
	}
    }
    return $heap
}

# Re-traverse bigcore's linked list, checking each chunk's address
# against the executable.  Don't use gdb_test_multiple as want only
# one pass/fail.  Don't use exp_continue as the regular expression
# involving $heap needs to be re-evaluated for each new response.

proc check_heap { dir heap } {
    global gdb_prompt
    set test "check ${dir} heap"
    set ok 1
    set lim 0
    send_gdb "print heap.${dir}\n"
    while { $ok } {
	gdb_expect {
	    -re " = \\(struct list \\*\\) [lindex $heap $lim].*$gdb_prompt $" {
		if { $lim >= [llength $heap] } {
		    pass "$test"
		    set ok 0
		} else {
		    incr lim
		    send_gdb "print \$.${dir}\n"
		}
	    }
	    -re ".*$gdb_prompt $" {
		fail "$test (address [lindex $heap $lim])"
		set ok 0
	    }
	    timeout {
		fail "$test (timeout)"
		set ok 0
	    }
	}
    }
}

# The bulk of the testcase.  DUMPER indicates who is supposed to dump
# the core.  It can be either "kernel", or "gdb".
proc test {dumper} {
    global binfile timeout corefile gdb_prompt

    # Run GDB on the bigcore program up-to where it will dump core.

    clean_restart ${binfile}
    gdb_test_no_output "set print sevenbit-strings"
    gdb_test_no_output "set width 0"

    # Get the core into the output directory.
    set_inferior_cwd_to_output_dir

    if {![runto_main]} {
	return 0
    }
    set print_core_line [gdb_get_line_number "Dump core"]
    gdb_test "tbreak $print_core_line"
    gdb_test continue ".*print_string.*"
    gdb_test next ".*0 = 0.*"

    set next_heap [extract_heap next]
    set prev_heap [extract_heap prev]

    # Save the total allocated size within GDB so that we can check
    # the core size later.
    gdb_test_no_output "set \$bytes_allocated = bytes_allocated" \
	"save heap size"

    # Now create a core dump.

    if {$dumper == "kernel"} {
	# Rename the core file to "TESTFILE.corefile.$dumper" rather
	# than just "core", to avoid problems with sys admin types
	# that like to regularly prune all files named "core" from the
	# system.

	# Some systems append "core" to the name of the program;
	# others append the name of the program to "core"; still
	# others (like Linux, as of May 2003) create cores named
	# "core.PID".

	# Save the process ID.  Some systems dump the core into
	# core.PID.
	set inferior_pid [get_inferior_pid]

	# Dump core using SIGABRT.
	set oldtimeout $timeout
	set timeout 600
	gdb_test "signal SIGABRT" "Program terminated with signal SIGABRT, .*"
	set timeout $oldtimeout

	# Find the corefile.
	set file [find_core_file $inferior_pid]
	if { $file != "" } {
	    remote_exec build "mv $file $corefile.$dumper"
	} else {
	    untested "can't generate a core file"
	    return 0
	}
    } elseif {$dumper == "gdb"} {
	gdb_gcore_cmd "$corefile.$dumper" "gcore corefile"
    } else {
	error "unhandled dumper: $dumper"
    }

    # Check that the corefile is plausibly large enough.  We're trying
    # to detect the case where the operating system has truncated the
    # file just before signed wraparound.  TCL, unfortunately, has a
    # similar problem - so use catch.  It can handle the "bad" size
    # but not necessarily the "good" one.  And we must use GDB for the
    # comparison, similarly.

    if {[catch {file size $corefile.$dumper} core_size] == 0} {
	set core_ok 0
	gdb_test_multiple "print \$bytes_allocated < $core_size" \
	    "check core size" {
	    -re " = 1\r\n$gdb_prompt $" {
		pass "check core size"
		set core_ok 1
	    }
	    -re " = 0\r\n$gdb_prompt $" {
		pass "check core size"
		set core_ok 0
	    }
	}
    } {
	# Probably failed due to the TCL build having problems with
	# very large values.  Since GDB uses a 64-bit off_t (when
	# possible) it shouldn't have this problem.  Assume that
	# things are going to work.  Without this assumption the test
	# is skiped on systems (such as i386 GNU/Linux with patched
	# kernel) which do pass.
	pass "check core size"
	set core_ok 1
    }
    if {! $core_ok} {
	untested "check core size (system does not support large corefiles)"
	return 0
    }

    # Now load up that core file.

    set test "load corefile"
    gdb_test_multiple "core $corefile.$dumper" "$test" {
	-re "A program is being debugged already.  Kill it. .y or n. " {
	    send_gdb "y\n"
	    exp_continue
	}
	-re "Core was generated by.*$gdb_prompt $" {
	    pass "$test"
	}
    }

    # Finally, re-traverse bigcore's linked list, checking each
    # chunk's address against the executable.

    check_heap next $next_heap
    check_heap prev $prev_heap
}

foreach_with_prefix dumper {kernel gdb} {
    # GDB's gcore is too slow when testing with the extended-gdbserver
    # board, since it requires reading all the inferior memory.
    if {$dumper == "gdb" && [target_info gdb_protocol] != ""} {
	continue
    }
    test $dumper
}