From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (qmail 98186 invoked by alias); 10 May 2017 11:48:02 -0000 Mailing-List: contact gdb-patches-help@sourceware.org; run by ezmlm Precedence: bulk List-Id: List-Subscribe: List-Archive: List-Post: List-Help: , Sender: gdb-patches-owner@sourceware.org Received: (qmail 97911 invoked by uid 89); 10 May 2017 11:48:01 -0000 Authentication-Results: sourceware.org; auth=none X-Virus-Found: No X-Spam-SWARE-Status: No, score=-24.7 required=5.0 tests=AWL,BAYES_00,GIT_PATCH_0,GIT_PATCH_1,GIT_PATCH_2,GIT_PATCH_3,KAM_LAZY_DOMAIN_SECURITY,RP_MATCHES_RCVD autolearn=ham version=3.3.2 spammy= X-HELO: mga11.intel.com Received: from mga11.intel.com (HELO mga11.intel.com) (192.55.52.93) by sourceware.org (qpsmtpd/0.93/v0.84-503-g423c35a) with ESMTP; Wed, 10 May 2017 11:47:57 +0000 Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by fmsmga102.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 10 May 2017 04:47:58 -0700 X-ExtLoop1: 1 Received: from irvmail001.ir.intel.com ([163.33.26.43]) by fmsmga001.fm.intel.com with ESMTP; 10 May 2017 04:47:57 -0700 Received: from ulvlx001.iul.intel.com (ulvlx001.iul.intel.com [172.28.207.17]) by irvmail001.ir.intel.com (8.14.3/8.13.6/MailSET/Hub) with ESMTP id v4ABlumF018146; Wed, 10 May 2017 12:47:56 +0100 Received: from ulvlx001.iul.intel.com (localhost [127.0.0.1]) by ulvlx001.iul.intel.com with ESMTP id v4ABlul7019904; Wed, 10 May 2017 13:47:56 +0200 Received: (from twiederh@localhost) by ulvlx001.iul.intel.com with œ id v4ABluDZ019900; Wed, 10 May 2017 13:47:56 +0200 From: Tim Wiederhake To: gdb-patches@sourceware.org Cc: markus.t.metzger@intel.com, simon.marchi@polymtl.ca Subject: [PATCH v4 08/12] btrace: Replace struct btrace_function::up. Date: Wed, 10 May 2017 11:48:00 -0000 Message-Id: <1494416867-19612-9-git-send-email-tim.wiederhake@intel.com> In-Reply-To: <1494416867-19612-1-git-send-email-tim.wiederhake@intel.com> References: <1494416867-19612-1-git-send-email-tim.wiederhake@intel.com> X-IsSubscribed: yes X-SW-Source: 2017-05/txt/msg00260.txt.bz2 This used to hold a function segment pointer. Change it to hold an index into the vector of function segments instead. 2017-05-10 Tim Wiederhake gdb/ChangeLog: * btrace.c (ftrace_find_call_by_number): New function. (ftrace_update_caller, ftrace_new_call, ftrace_new_tailcall, ftrace_get_caller, ftrace_find_call, ftrace_new_return, ftrace_match_backtrace, ftrace_connect_bfun, ftrace_connect_backtrace, ftrace_bridge_gap, btrace_bridge_gaps): Use btrace_function::up as an index. * btrace.h (struct btrace_function): Turn UP into an index. * python/py-record-btrace.c (btpy_call_up): Use btrace_function::up as an index. * record-btrace.c (record_btrace_frame_unwind_stop_reason, record_btrace_frame_prev_register, record_btrace_frame_sniffer, record_btrace_tailcall_frame_sniffe): Same. --- gdb/btrace.c | 141 ++++++++++++++++++++++++++---------------- gdb/btrace.h | 6 +- gdb/python/py-record-btrace.c | 4 +- gdb/record-btrace.c | 18 +++--- 4 files changed, 106 insertions(+), 63 deletions(-) diff --git a/gdb/btrace.c b/gdb/btrace.c index 8c86a4f..f924dda 100644 --- a/gdb/btrace.c +++ b/gdb/btrace.c @@ -156,6 +156,19 @@ ftrace_call_num_insn (const struct btrace_function* bfun) return VEC_length (btrace_insn_s, bfun->insn); } +/* Return the function segment with the given NUMBER or NULL if no such segment + exists. BTINFO is the branch trace information for the current thread. */ + +static struct btrace_function * +ftrace_find_call_by_number (const struct btrace_thread_info *btinfo, + unsigned int number) +{ + if (number == 0 || number > btinfo->functions.size ()) + return NULL; + + return btinfo->functions[number - 1]; +} + /* Return non-zero if BFUN does not match MFUN and FUN, return zero otherwise. */ @@ -249,10 +262,10 @@ ftrace_update_caller (struct btrace_function *bfun, struct btrace_function *caller, enum btrace_function_flag flags) { - if (bfun->up != NULL) + if (bfun->up != 0) ftrace_debug (bfun, "updating caller"); - bfun->up = caller; + bfun->up = caller->number; bfun->flags = flags; ftrace_debug (bfun, "set caller"); @@ -290,8 +303,7 @@ ftrace_new_call (struct btrace_thread_info *btinfo, const unsigned int length = btinfo->functions.size (); struct btrace_function *bfun = ftrace_new_function (btinfo, mfun, fun); - if (length != 0) - bfun->up = btinfo->functions[length - 1]; + bfun->up = length; bfun->level += 1; ftrace_debug (bfun, "new call"); @@ -311,8 +323,7 @@ ftrace_new_tailcall (struct btrace_thread_info *btinfo, const unsigned int length = btinfo->functions.size (); struct btrace_function *bfun = ftrace_new_function (btinfo, mfun, fun); - if (length != 0) - bfun->up = btinfo->functions[length - 1]; + bfun->up = length; bfun->level += 1; bfun->flags |= BFUN_UP_LINKS_TO_TAILCALL; @@ -322,26 +333,30 @@ ftrace_new_tailcall (struct btrace_thread_info *btinfo, } /* Return the caller of BFUN or NULL if there is none. This function skips - tail calls in the call chain. */ + tail calls in the call chain. BTINFO is the branch trace information for + the current thread. */ static struct btrace_function * -ftrace_get_caller (struct btrace_function *bfun) +ftrace_get_caller (struct btrace_thread_info *btinfo, + struct btrace_function *bfun) { - for (; bfun != NULL; bfun = bfun->up) + for (; bfun != NULL; bfun = ftrace_find_call_by_number (btinfo, bfun->up)) if ((bfun->flags & BFUN_UP_LINKS_TO_TAILCALL) == 0) - return bfun->up; + return ftrace_find_call_by_number (btinfo, bfun->up); return NULL; } /* Find the innermost caller in the back trace of BFUN with MFUN/FUN - symbol information. */ + symbol information. BTINFO is the branch trace information for the current + thread. */ static struct btrace_function * -ftrace_find_caller (struct btrace_function *bfun, +ftrace_find_caller (struct btrace_thread_info *btinfo, + struct btrace_function *bfun, struct minimal_symbol *mfun, struct symbol *fun) { - for (; bfun != NULL; bfun = bfun->up) + for (; bfun != NULL; bfun = ftrace_find_call_by_number (btinfo, bfun->up)) { /* Skip functions with incompatible symbol information. */ if (ftrace_function_switched (bfun, mfun, fun)) @@ -356,12 +371,14 @@ ftrace_find_caller (struct btrace_function *bfun, /* Find the innermost caller in the back trace of BFUN, skipping all function segments that do not end with a call instruction (e.g. - tail calls ending with a jump). */ + tail calls ending with a jump). BTINFO is the branch trace information for + the current thread. */ static struct btrace_function * -ftrace_find_call (struct btrace_function *bfun) +ftrace_find_call (struct btrace_thread_info *btinfo, + struct btrace_function *bfun) { - for (; bfun != NULL; bfun = bfun->up) + for (; bfun != NULL; bfun = ftrace_find_call_by_number (btinfo, bfun->up)) { struct btrace_insn *last; @@ -395,7 +412,8 @@ ftrace_new_return (struct btrace_thread_info *btinfo, /* It is important to start at PREV's caller. Otherwise, we might find PREV itself, if PREV is a recursive function. */ - caller = ftrace_find_caller (prev->up, mfun, fun); + caller = ftrace_find_call_by_number (btinfo, prev->up); + caller = ftrace_find_caller (btinfo, caller, mfun, fun); if (caller != NULL) { /* The caller of PREV is the preceding btrace function segment in this @@ -420,7 +438,8 @@ ftrace_new_return (struct btrace_thread_info *btinfo, wrong or that the call is simply not included in the trace. */ /* Let's search for some actual call. */ - caller = ftrace_find_call (prev->up); + caller = ftrace_find_call_by_number (btinfo, prev->up); + caller = ftrace_find_call (btinfo, caller); if (caller == NULL) { /* There is no call in PREV's back trace. We assume that the @@ -428,8 +447,8 @@ ftrace_new_return (struct btrace_thread_info *btinfo, /* Let's find the topmost function and add a new caller for it. This should handle a series of initial tail calls. */ - while (prev->up != NULL) - prev = prev->up; + while (prev->up != 0) + prev = ftrace_find_call_by_number (btinfo, prev->up); bfun->level = prev->level - 1; @@ -449,7 +468,7 @@ ftrace_new_return (struct btrace_thread_info *btinfo, on the same level as they are. This should handle things like schedule () correctly where we're switching contexts. */ - prev->up = bfun; + prev->up = bfun->number; prev->flags = BFUN_UP_LINKS_TO_RET; ftrace_debug (bfun, "new return - unknown caller"); @@ -653,10 +672,11 @@ ftrace_classify_insn (struct gdbarch *gdbarch, CORE_ADDR pc) /* Try to match the back trace at LHS to the back trace at RHS. Returns the number of matching function segments or zero if the back traces do not - match. */ + match. BTINFO is the branch trace information for the current thread. */ static int -ftrace_match_backtrace (struct btrace_function *lhs, +ftrace_match_backtrace (struct btrace_thread_info *btinfo, + struct btrace_function *lhs, struct btrace_function *rhs) { int matches; @@ -666,8 +686,8 @@ ftrace_match_backtrace (struct btrace_function *lhs, if (ftrace_function_switched (lhs, rhs->msym, rhs->sym)) return 0; - lhs = ftrace_get_caller (lhs); - rhs = ftrace_get_caller (rhs); + lhs = ftrace_get_caller (btinfo, lhs); + rhs = ftrace_get_caller (btinfo, rhs); } return matches; @@ -721,10 +741,12 @@ ftrace_compute_global_level_offset (struct btrace_thread_info *btinfo) } /* Connect the function segments PREV and NEXT in a bottom-to-top walk as in - ftrace_connect_backtrace. */ + ftrace_connect_backtrace. BTINFO is the branch trace information for the + current thread. */ static void -ftrace_connect_bfun (struct btrace_function *prev, +ftrace_connect_bfun (struct btrace_thread_info *btinfo, + struct btrace_function *prev, struct btrace_function *next) { DEBUG_FTRACE ("connecting..."); @@ -742,20 +764,26 @@ ftrace_connect_bfun (struct btrace_function *prev, ftrace_fixup_level (next, prev->level - next->level); /* If we run out of back trace for one, let's use the other's. */ - if (prev->up == NULL) + if (prev->up == 0) { - if (next->up != NULL) + const btrace_function_flags flags = next->flags; + + next = ftrace_find_call_by_number (btinfo, next->up); + if (next != NULL) { DEBUG_FTRACE ("using next's callers"); - ftrace_fixup_caller (prev, next->up, next->flags); + ftrace_fixup_caller (prev, next, flags); } } - else if (next->up == NULL) + else if (next->up == 0) { - if (prev->up != NULL) + const btrace_function_flags flags = prev->flags; + + prev = ftrace_find_call_by_number (btinfo, prev->up); + if (prev != NULL) { DEBUG_FTRACE ("using prev's callers"); - ftrace_fixup_caller (next, prev->up, prev->flags); + ftrace_fixup_caller (next, prev, flags); } } else @@ -773,26 +801,29 @@ ftrace_connect_bfun (struct btrace_function *prev, if ((prev->flags & BFUN_UP_LINKS_TO_TAILCALL) != 0) { struct btrace_function *caller; - btrace_function_flags flags; + btrace_function_flags next_flags, prev_flags; /* We checked NEXT->UP above so CALLER can't be NULL. */ - caller = next->up; - flags = next->flags; + caller = ftrace_find_call_by_number (btinfo, next->up); + next_flags = next->flags; + prev_flags = prev->flags; DEBUG_FTRACE ("adding prev's tail calls to next"); - ftrace_fixup_caller (next, prev->up, prev->flags); + prev = ftrace_find_call_by_number (btinfo, prev->up); + ftrace_fixup_caller (next, prev, prev_flags); - for (prev = prev->up; prev != NULL; prev = prev->up) + for (; prev != NULL; prev = ftrace_find_call_by_number (btinfo, + prev->up)) { /* At the end of PREV's back trace, continue with CALLER. */ - if (prev->up == NULL) + if (prev->up == 0) { DEBUG_FTRACE ("fixing up link for tailcall chain"); ftrace_debug (prev, "..top"); ftrace_debug (caller, "..up"); - ftrace_fixup_caller (prev, caller, flags); + ftrace_fixup_caller (prev, caller, next_flags); /* If we skipped any tail calls, this may move CALLER to a different function level. @@ -820,10 +851,12 @@ ftrace_connect_bfun (struct btrace_function *prev, /* Connect function segments on the same level in the back trace at LHS and RHS. The back traces at LHS and RHS are expected to match according to - ftrace_match_backtrace. */ + ftrace_match_backtrace. BTINFO is the branch trace information for the + current thread. */ static void -ftrace_connect_backtrace (struct btrace_function *lhs, +ftrace_connect_backtrace (struct btrace_thread_info *btinfo, + struct btrace_function *lhs, struct btrace_function *rhs) { while (lhs != NULL && rhs != NULL) @@ -836,20 +869,22 @@ ftrace_connect_backtrace (struct btrace_function *lhs, prev = lhs; next = rhs; - lhs = ftrace_get_caller (lhs); - rhs = ftrace_get_caller (rhs); + lhs = ftrace_get_caller (btinfo, lhs); + rhs = ftrace_get_caller (btinfo, rhs); - ftrace_connect_bfun (prev, next); + ftrace_connect_bfun (btinfo, prev, next); } } /* Bridge the gap between two function segments left and right of a gap if their - respective back traces match in at least MIN_MATCHES functions. + respective back traces match in at least MIN_MATCHES functions. BTINFO is + the branch trace information for the current thread. Returns non-zero if the gap could be bridged, zero otherwise. */ static int -ftrace_bridge_gap (struct btrace_function *lhs, struct btrace_function *rhs, +ftrace_bridge_gap (struct btrace_thread_info *btinfo, + struct btrace_function *lhs, struct btrace_function *rhs, int min_matches) { struct btrace_function *best_l, *best_r, *cand_l, *cand_r; @@ -865,12 +900,14 @@ ftrace_bridge_gap (struct btrace_function *lhs, struct btrace_function *rhs, /* We search the back traces of LHS and RHS for valid connections and connect the two functon segments that give the longest combined back trace. */ - for (cand_l = lhs; cand_l != NULL; cand_l = ftrace_get_caller (cand_l)) - for (cand_r = rhs; cand_r != NULL; cand_r = ftrace_get_caller (cand_r)) + for (cand_l = lhs; cand_l != NULL; + cand_l = ftrace_get_caller (btinfo, cand_l)) + for (cand_r = rhs; cand_r != NULL; + cand_r = ftrace_get_caller (btinfo, cand_r)) { int matches; - matches = ftrace_match_backtrace (cand_l, cand_r); + matches = ftrace_match_backtrace (btinfo, cand_l, cand_r); if (best_matches < matches) { best_matches = matches; @@ -897,7 +934,7 @@ ftrace_bridge_gap (struct btrace_function *lhs, struct btrace_function *rhs, BEST_L to BEST_R as they will already be on the same level. */ ftrace_fixup_level (rhs, best_l->level - best_r->level); - ftrace_connect_backtrace (best_l, best_r); + ftrace_connect_backtrace (btinfo, best_l, best_r); return best_matches; } @@ -955,7 +992,7 @@ btrace_bridge_gaps (struct thread_info *tp, VEC (bfun_s) **gaps) if (rhs == NULL) continue; - bridged = ftrace_bridge_gap (lhs, rhs, min_matches); + bridged = ftrace_bridge_gap (&tp->btrace, lhs, rhs, min_matches); /* Keep track of gaps we were not able to bridge and try again. If we just pushed them to the end of GAPS we would risk an diff --git a/gdb/btrace.h b/gdb/btrace.h index 601b736..3d6e4e1 100644 --- a/gdb/btrace.h +++ b/gdb/btrace.h @@ -154,8 +154,10 @@ struct btrace_function /* The previous and next function in control flow order. */ struct btrace_func_link flow; - /* The directly preceding function segment in a (fake) call stack. */ - struct btrace_function *up; + /* The function segment number of the directly preceding function segment in + a (fake) call stack. Will be zero if there is no such function segment in + the record. */ + unsigned int up; /* The instructions in this function segment. The instruction vector will be empty if the function segment diff --git a/gdb/python/py-record-btrace.c b/gdb/python/py-record-btrace.c index d684561..9dd2199 100644 --- a/gdb/python/py-record-btrace.c +++ b/gdb/python/py-record-btrace.c @@ -398,11 +398,11 @@ recpy_bt_func_up (PyObject *self, void *closure) if (func == NULL) return NULL; - if (func->up == NULL) + if (func->up == 0) Py_RETURN_NONE; return recpy_func_new (((recpy_element_object *) self)->ptid, - RECORD_METHOD_BTRACE, func->up->number); + RECORD_METHOD_BTRACE, func->up); } /* Implementation of RecordFunctionSegment.prev [RecordFunctionSegment] for diff --git a/gdb/record-btrace.c b/gdb/record-btrace.c index 5c230e7..a27521c 100644 --- a/gdb/record-btrace.c +++ b/gdb/record-btrace.c @@ -1570,7 +1570,7 @@ record_btrace_frame_unwind_stop_reason (struct frame_info *this_frame, bfun = cache->bfun; gdb_assert (bfun != NULL); - if (bfun->up == NULL) + if (bfun->up == 0) return UNWIND_UNAVAILABLE; return UNWIND_NO_REASON; @@ -1629,11 +1629,12 @@ record_btrace_frame_prev_register (struct frame_info *this_frame, bfun = cache->bfun; gdb_assert (bfun != NULL); - caller = bfun->up; - if (caller == NULL) + if (bfun->up == 0) throw_error (NOT_AVAILABLE_ERROR, _("No caller in btrace record history")); + caller = cache->tp->btrace.functions[bfun->up - 1]; + if ((bfun->flags & BFUN_UP_LINKS_TO_RET) != 0) { insn = VEC_index (btrace_insn_s, caller->insn, 0); @@ -1686,7 +1687,7 @@ record_btrace_frame_sniffer (const struct frame_unwind *self, callee = btrace_get_frame_function (next); if (callee != NULL && (callee->flags & BFUN_UP_LINKS_TO_TAILCALL) == 0) - bfun = callee->up; + bfun = tp->btrace.functions[callee->up - 1]; } if (bfun == NULL) @@ -1713,6 +1714,7 @@ record_btrace_tailcall_frame_sniffer (const struct frame_unwind *self, { const struct btrace_function *bfun, *callee; struct btrace_frame_cache *cache; + struct thread_info *tinfo; struct frame_info *next; next = get_next_frame (this_frame); @@ -1726,16 +1728,18 @@ record_btrace_tailcall_frame_sniffer (const struct frame_unwind *self, if ((callee->flags & BFUN_UP_LINKS_TO_TAILCALL) == 0) return 0; - bfun = callee->up; - if (bfun == NULL) + if (callee->up == 0) return 0; + tinfo = find_thread_ptid (inferior_ptid); + bfun = tinfo->btrace.functions[callee->up - 1]; + DEBUG ("[frame] sniffed tailcall frame for %s on level %d", btrace_get_bfun_name (bfun), bfun->level); /* This is our frame. Initialize the frame cache. */ cache = bfcache_new (this_frame); - cache->tp = find_thread_ptid (inferior_ptid); + cache->tp = tinfo; cache->bfun = bfun; *this_cache = cache; -- 2.7.4