use rustc_data_structures::fx::FxHashSet; use rustc_middle::mir; use rustc_middle::ty::TyCtxt; use rustc_span::source_map::SourceMap; use rustc_span::{BytePos, DesugaringKind, ExpnKind, MacroKind, Span}; use tracing::instrument; use crate::coverage::graph::{BasicCoverageBlock, CoverageGraph}; use crate::coverage::spans::from_mir::{Hole, RawSpanFromMir, SpanFromMir}; use crate::coverage::{ExtractedHirInfo, mappings, unexpand}; mod from_mir; pub(super) fn extract_refined_covspans<'tcx>( tcx: TyCtxt<'tcx>, mir_body: &mir::Body<'tcx>, hir_info: &ExtractedHirInfo, graph: &CoverageGraph, code_mappings: &mut impl Extend, ) { let &ExtractedHirInfo { body_span, .. } = hir_info; let raw_spans = from_mir::extract_raw_spans_from_mir(mir_body, graph); let mut covspans = raw_spans .into_iter() .filter_map(|RawSpanFromMir { raw_span, bcb }| try { let (span, expn_kind) = unexpand::unexpand_into_body_span_with_expn_kind(raw_span, body_span)?; // Discard any spans that fill the entire body, because they tend // to represent compiler-inserted code, e.g. implicitly returning `()`. if span.source_equal(body_span) { return None; }; SpanFromMir { span, expn_kind, bcb } }) .collect::>(); // Only proceed if we found at least one usable span. if covspans.is_empty() { return; } // Also add the function signature span, if available. // Otherwise, add a fake span at the start of the body, to avoid an ugly // gap between the start of the body and the first real span. // FIXME: Find a more principled way to solve this problem. covspans.push(SpanFromMir::for_fn_sig( hir_info.fn_sig_span.unwrap_or_else(|| body_span.shrink_to_lo()), )); // First, perform the passes that need macro information. covspans.sort_by(|a, b| graph.cmp_in_dominator_order(a.bcb, b.bcb)); remove_unwanted_expansion_spans(&mut covspans); shrink_visible_macro_spans(tcx, &mut covspans); // We no longer need the extra information in `SpanFromMir`, so convert to `Covspan`. let mut covspans = covspans.into_iter().map(SpanFromMir::into_covspan).collect::>(); let compare_covspans = |a: &Covspan, b: &Covspan| { compare_spans(a.span, b.span) // After deduplication, we want to keep only the most-dominated BCB. .then_with(|| graph.cmp_in_dominator_order(a.bcb, b.bcb).reverse()) }; covspans.sort_by(compare_covspans); // Among covspans with the same span, keep only one, // preferring the one with the most-dominated BCB. // (Ideally we should try to preserve _all_ non-dominating BCBs, but that // requires a lot more complexity in the span refiner, for little benefit.) covspans.dedup_by(|b, a| a.span.source_equal(b.span)); // Sort the holes, and merge overlapping/adjacent holes. let mut holes = hir_info .hole_spans .iter() .copied() // Discard any holes that aren't directly visible within the body span. .filter(|&hole_span| body_span.contains(hole_span) && body_span.eq_ctxt(hole_span)) .map(|span| Hole { span }) .collect::>(); holes.sort_by(|a, b| compare_spans(a.span, b.span)); holes.dedup_by(|b, a| a.merge_if_overlapping_or_adjacent(b)); // Discard any span that overlaps with a hole. discard_spans_overlapping_holes(&mut covspans, &holes); // Discard spans that overlap in unwanted ways. let mut covspans = remove_unwanted_overlapping_spans(covspans); // For all empty spans, either enlarge them to be non-empty, or discard them. let source_map = tcx.sess.source_map(); covspans.retain_mut(|covspan| { let Some(span) = ensure_non_empty_span(source_map, covspan.span) else { return false }; covspan.span = span; true }); // Merge covspans that can be merged. covspans.dedup_by(|b, a| a.merge_if_eligible(b)); code_mappings.extend(covspans.into_iter().map(|Covspan { span, bcb }| { // Each span produced by the refiner represents an ordinary code region. mappings::CodeMapping { span, bcb } })); } /// Macros that expand into branches (e.g. `assert!`, `trace!`) tend to generate /// multiple condition/consequent blocks that have the span of the whole macro /// invocation, which is unhelpful. Keeping only the first such span seems to /// give better mappings, so remove the others. /// /// Similarly, `await` expands to a branch on the discriminant of `Poll`, which /// leads to incorrect coverage if the `Future` is immediately ready (#98712). /// /// (The input spans should be sorted in BCB dominator order, so that the /// retained "first" span is likely to dominate the others.) fn remove_unwanted_expansion_spans(covspans: &mut Vec) { let mut deduplicated_spans = FxHashSet::default(); covspans.retain(|covspan| { match covspan.expn_kind { // Retain only the first await-related or macro-expanded covspan with this span. Some(ExpnKind::Desugaring(DesugaringKind::Await)) => { deduplicated_spans.insert(covspan.span) } Some(ExpnKind::Macro(MacroKind::Bang, _)) => deduplicated_spans.insert(covspan.span), // Ignore (retain) other spans. _ => true, } }); } /// When a span corresponds to a macro invocation that is visible from the /// function body, truncate it to just the macro name plus `!`. /// This seems to give better results for code that uses macros. fn shrink_visible_macro_spans(tcx: TyCtxt<'_>, covspans: &mut Vec) { let source_map = tcx.sess.source_map(); for covspan in covspans { if matches!(covspan.expn_kind, Some(ExpnKind::Macro(MacroKind::Bang, _))) { covspan.span = source_map.span_through_char(covspan.span, '!'); } } } /// Discard all covspans that overlap a hole. /// /// The lists of covspans and holes must be sorted, and any holes that overlap /// with each other must have already been merged. fn discard_spans_overlapping_holes(covspans: &mut Vec, holes: &[Hole]) { debug_assert!(covspans.is_sorted_by(|a, b| compare_spans(a.span, b.span).is_le())); debug_assert!(holes.is_sorted_by(|a, b| compare_spans(a.span, b.span).is_le())); debug_assert!(holes.array_windows().all(|[a, b]| !a.span.overlaps_or_adjacent(b.span))); let mut curr_hole = 0usize; let mut overlaps_hole = |covspan: &Covspan| -> bool { while let Some(hole) = holes.get(curr_hole) { // Both lists are sorted, so we can permanently skip any holes that // end before the start of the current span. if hole.span.hi() <= covspan.span.lo() { curr_hole += 1; continue; } return hole.span.overlaps(covspan.span); } // No holes left, so this covspan doesn't overlap with any holes. false }; covspans.retain(|covspan| !overlaps_hole(covspan)); } /// Takes a list of sorted spans extracted from MIR, and "refines" /// those spans by removing spans that overlap in unwanted ways. #[instrument(level = "debug")] fn remove_unwanted_overlapping_spans(sorted_spans: Vec) -> Vec { debug_assert!(sorted_spans.is_sorted_by(|a, b| compare_spans(a.span, b.span).is_le())); // Holds spans that have been read from the input vector, but haven't yet // been committed to the output vector. let mut pending = vec![]; let mut refined = vec![]; for curr in sorted_spans { pending.retain(|prev: &Covspan| { if prev.span.hi() <= curr.span.lo() { // There's no overlap between the previous/current covspans, // so move the previous one into the refined list. refined.push(prev.clone()); false } else { // Otherwise, retain the previous covspan only if it has the // same BCB. This tends to discard long outer spans that enclose // smaller inner spans with different control flow. prev.bcb == curr.bcb } }); pending.push(curr); } // Drain the rest of the pending list into the refined list. refined.extend(pending); refined } #[derive(Clone, Debug)] struct Covspan { span: Span, bcb: BasicCoverageBlock, } impl Covspan { /// If `self` and `other` can be merged, mutates `self.span` to also /// include `other.span` and returns true. /// /// Two covspans can be merged if they have the same BCB, and they are /// overlapping or adjacent. fn merge_if_eligible(&mut self, other: &Self) -> bool { let eligible_for_merge = |a: &Self, b: &Self| (a.bcb == b.bcb) && a.span.overlaps_or_adjacent(b.span); if eligible_for_merge(self, other) { self.span = self.span.to(other.span); true } else { false } } } /// Compares two spans in (lo ascending, hi descending) order. fn compare_spans(a: Span, b: Span) -> std::cmp::Ordering { // First sort by span start. Ord::cmp(&a.lo(), &b.lo()) // If span starts are the same, sort by span end in reverse order. // This ensures that if spans A and B are adjacent in the list, // and they overlap but are not equal, then either: // - Span A extends further left, or // - Both have the same start and span A extends further right .then_with(|| Ord::cmp(&a.hi(), &b.hi()).reverse()) } fn ensure_non_empty_span(source_map: &SourceMap, span: Span) -> Option { if !span.is_empty() { return Some(span); } // The span is empty, so try to enlarge it to cover an adjacent '{' or '}'. source_map .span_to_source(span, |src, start, end| try { // Adjusting span endpoints by `BytePos(1)` is normally a bug, // but in this case we have specifically checked that the character // we're skipping over is one of two specific ASCII characters, so // adjusting by exactly 1 byte is correct. if src.as_bytes().get(end).copied() == Some(b'{') { Some(span.with_hi(span.hi() + BytePos(1))) } else if start > 0 && src.as_bytes()[start - 1] == b'}' { Some(span.with_lo(span.lo() - BytePos(1))) } else { None } }) .ok()? }