From: "Schimpe, Christina" <christina.schimpe@intel.com>
To: Kevin Buettner <kevinb@redhat.com>,
"gdb-patches@sourceware.org" <gdb-patches@sourceware.org>
Subject: RE: [PATCH v2] gcore: Handle unreadable pages within readable memory regions
Date: Tue, 17 Feb 2026 17:44:06 +0000 [thread overview]
Message-ID: <SN7PR11MB76383EB42B11B8207E8D7484F96DA@SN7PR11MB7638.namprd11.prod.outlook.com> (raw)
In-Reply-To: <20260212194039.1717054-1-kevinb@redhat.com>
> -----Original Message-----
> From: Kevin Buettner <kevinb@redhat.com>
> Sent: Donnerstag, 12. Februar 2026 20:40
> To: gdb-patches@sourceware.org
> Cc: Kevin Buettner <kevinb@redhat.com>
> Subject: [PATCH v2] gcore: Handle unreadable pages within readable memory
> regions
>
> GLIBC 2.42 changed how thread stack guard pages are implemented [2].
> In GLIBC 2.41 and earlier, guard pages were set up using mprotect() to mark
> guard regions with no permissions. Once configured, guard pages were
> visible as separate entries in /proc/PID/maps with no permissions (i.e. they're
> inaccessible). In GLIBC 2.42, guard pages are installed using the kernel's
> MADV_GUARD_INSTALL mechanism [1], which marks them at the page table
> entry (PTE) level within the existing mapping.
>
> As a consequence, guard pages do not appear as separate entries in
> /proc/PID/maps, but remain as part of the containing mapping. Moreover,
> thread stacks from multiple mmap() calls may be merged into a single virtual
> memory area (VMA) with read and write permissions since there's no guard
> page VMA to separate them. These guard pages cannot be distinguished by
> examining VMA listings but do return EIO when read from /proc/PID/mem.
>
> GDB's gcore code reads /proc/PID/smaps to discover memory regions and
> creates one BFD section per mapping. (On linux, this is performed in
> linux_find_memory_regions_full in linux-tdep.c.) With the old layout,
> memory areas with guard pages appeared separately with no permissions,
> which were filtered out. Each thread stack became its own section
> containing only readable data. With the new layout, using
> MADV_GUARD_INSTALL instead of the older mechanism, it's often the case
> that thread stacks created with multiple calls to mmap() are exposed as a
> single mapping appearing in /proc/PID/smaps with read and write
> permissions. Should that happen, GDB's code creates a single section
> covering all thread stacks and their guard pages. (Even if each thread stack
> appears in its own mapping, the fact remains that there will be an
> inaccessible portion of the mapping. When one or more thread stacks are
> coalesced into a single mapping, there will be several inaccessible "holes"
> representing the guard pages.)
>
> When gcore_copy_callback copies section contents, it reads memory in 1MB
> (MAX_COPY_BYTES) chunks. If any page in the chunk is a guard page, the call
> to target_read_memory() fails. The old code responded by breaking out of
> the copy loop, abandoning the entire section. This prevents correct copying
> of thread stack data, resulting in core files with zero-filled thread stacks,
> resulting in nearly empty backtraces.
>
> Fix this by falling back to page-by-page reading when a 1MB chunk read fails.
> Individual pages that cannot be read are filled with zeros, allowing the
> remaining readable memory to be captured.
>
> I also considered a simpler change using the value of FALLBACK_PAGE_SIZE
> (4096) as the read size instead of MAX_COPY_BYTES (1MB). This would avoid
> the fallback logic but would cause up to 256x more syscalls. The proposed
> approach also allows meaningful warnings:
> we warn only if an entire region is unreadable (indicating a real problem),
> whereas per-page reads would make it harder to distinguish guard page
> failures from actual errors. Since guard pages are at offset 0 for downward-
> growing stacks, a large target_read_memory() fails early at the first
> unreadable byte anyway.
>
> With this fix, I see 16 failures resolved in the following test cases:
>
> gdb.ada/task_switch_in_core.exp
> gdb.arch/i386-tls-regs.exp
> gdb.threads/threadcrash.exp
> gdb.threads/tls-core.exp
>
> Looking at just one of these, from gdb.log without the fix, I see:
>
> thread apply 5 backtrace
>
> Thread 5 (LWP 3414829):
> #0 0x00007ffff7d1d982 in __syscall_cancel_arch () from /lib64/libc.so.6
> #1 0x0000000000000000 in ?? ()
> (gdb) FAIL: gdb.threads/threadcrash.exp: test_gcore: thread apply 5
> backtrace
>
> And this is what it looks like with the fix in place (some paths have been
> shortened):
>
> thread apply 5 backtrace
>
> Thread 5 (Thread 0x7fffeffff6c0 (LWP 1282651) "threadcrash"):
> #0 0x00007ffff7d1d982 in __syscall_cancel_arch () from /lib64/libc.so.6
> #1 0x00007ffff7d11c3c in __internal_syscall_cancel () from /lib64/libc.so.6
> #2 0x00007ffff7d61b62 in clock_nanosleep@GLIBC_2.2.5 () from
> /lib64/libc.so.6
> #3 0x00007ffff7d6db37 in nanosleep () from /lib64/libc.so.6
> #4 0x00007ffff7d8008e in sleep () from /lib64/libc.so.6
> #5 0x00000000004006a8 in do_syscall_task (location=NORMAL) at
> threadcrash.c:158
> #6 0x0000000000400885 in thread_function (arg=0x404340) at
> threadcrash.c:277
> #7 0x00007ffff7d15464 in start_thread () from /lib64/libc.so.6
> #8 0x00007ffff7d985ac in __clone3 () from /lib64/libc.so.6
> (gdb) PASS: gdb.threads/threadcrash.exp: test_live_inferior: thread apply 5
> backtrace
>
> Regression testing on Fedora 42 (glibc 2.41) shows no new failures.
>
> The v1 patch used SPARSE_BLOCK_SIZE as the fallback size. While it was the
> correct size, it's used for an entirely different purpose elsewhere in this file.
> This v2 commit introduces the constant FALLBACK_PAGE_SIZE instead.
>
> References:
>
> [1] Linux commit 662df3e5c376 ("mm: madvise: implement lightweight
> guard page mechanism")
>
> https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?i
> d=662df3e5c37666d6ed75c88098699e070a4b35b5
> [2] glibc commit a6fbe36b7f31 ("nptl: Add support for setup guard
> pages with MADV_GUARD_INSTALL")
>
> https://sourceware.org/git/?p=glibc.git;a=commit;h=a6fbe36b7f3129298142
> 2692236465ab56670ea9
>
> Claude Opus 4.5 and GLM 4.7 assisted with the development of this commit.
For binutils there is a guideline for LLM generated content:
https://sourceware.org/binutils/wiki/LLM_Generated_Content
Do we have something similar for GDB ?
Thanks,
Christina
Intel Deutschland GmbH
Registered Address: Dornacher Straße 1, 85622 Feldkirchen, Germany
Tel: +49 89 991 430, www.intel.de
Managing Directors: Harry Demas, Jeffrey Schneiderman, Yin Chong Sorrell
Chairperson of the Supervisory Board: Nicole Lau
Registered Seat: Munich
Commercial Register: Amtsgericht München HRB 186928
next prev parent reply other threads:[~2026-02-17 17:44 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-02-12 19:39 Kevin Buettner
2026-02-17 14:19 ` Tom de Vries
2026-02-17 17:50 ` Kevin Buettner
2026-02-17 17:44 ` Schimpe, Christina [this message]
2026-02-17 18:16 ` Kevin Buettner
2026-02-17 21:22 ` Schimpe, Christina
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=SN7PR11MB76383EB42B11B8207E8D7484F96DA@SN7PR11MB7638.namprd11.prod.outlook.com \
--to=christina.schimpe@intel.com \
--cc=gdb-patches@sourceware.org \
--cc=kevinb@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox