Skip to content

Commit 32e6cbb

Browse files
committed
Rollup merge of rust-lang#33566 - dotdash:biased_switch, r=nagisa
[MIR trans] Optimize trans for biased switches Currently, all switches in MIR are exhausitive, meaning that we can have a lot of arms that all go to the same basic block, the extreme case being an if-let expression which results in just 2 possible cases, be might end up with hundreds of arms for large enums. To improve this situation and give LLVM less code to chew on, we can detect whether there's a pre-dominant target basic block in a switch and then promote this to be the default target, not translating the corresponding arms at all. In combination with rust-lang#33544 this makes unoptimized MIR trans of nickel.rs as fast as using old trans and greatly improves the times for optimized builds, which are only 30-40% slower instead of ~300%. cc rust-lang#33111
2 parents 5203448 + 49b2cdf commit 32e6cbb

File tree

1 file changed

+26
-10
lines changed

1 file changed

+26
-10
lines changed

src/librustc_trans/mir/block.rs

+26-10
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@ use meth;
2424
use type_of;
2525
use glue;
2626
use type_::Type;
27+
use rustc_data_structures::fnv::FnvHashMap;
2728

2829
use super::{MirContext, TempRef, drop};
2930
use super::constant::Const;
@@ -95,17 +96,32 @@ impl<'bcx, 'tcx> MirContext<'bcx, 'tcx> {
9596
adt::trans_get_discr(bcx, &repr, discr_lvalue.llval, None, true)
9697
);
9798

98-
// The else branch of the Switch can't be hit, so branch to an unreachable
99-
// instruction so LLVM knows that
100-
let unreachable_blk = self.unreachable_block();
101-
let switch = bcx.switch(discr, unreachable_blk.llbb, targets.len());
99+
let mut bb_hist = FnvHashMap();
100+
for target in targets {
101+
*bb_hist.entry(target).or_insert(0) += 1;
102+
}
103+
let (default_bb, default_blk) = match bb_hist.iter().max_by_key(|&(_, c)| c) {
104+
// If a single target basic blocks is predominant, promote that to be the
105+
// default case for the switch instruction to reduce the size of the generated
106+
// code. This is especially helpful in cases like an if-let on a huge enum.
107+
// Note: This optimization is only valid for exhaustive matches.
108+
Some((&&bb, &c)) if c > targets.len() / 2 => {
109+
(Some(bb), self.blocks[bb.index()])
110+
}
111+
// We're generating an exhaustive switch, so the else branch
112+
// can't be hit. Branching to an unreachable instruction
113+
// lets LLVM know this
114+
_ => (None, self.unreachable_block())
115+
};
116+
let switch = bcx.switch(discr, default_blk.llbb, targets.len());
102117
assert_eq!(adt_def.variants.len(), targets.len());
103-
for (adt_variant, target) in adt_def.variants.iter().zip(targets) {
104-
let llval = bcx.with_block(|bcx|
105-
adt::trans_case(bcx, &repr, Disr::from(adt_variant.disr_val))
106-
);
107-
let llbb = self.llblock(*target);
108-
build::AddCase(switch, llval, llbb)
118+
for (adt_variant, &target) in adt_def.variants.iter().zip(targets) {
119+
if default_bb != Some(target) {
120+
let llbb = self.llblock(target);
121+
let llval = bcx.with_block(|bcx| adt::trans_case(
122+
bcx, &repr, Disr::from(adt_variant.disr_val)));
123+
build::AddCase(switch, llval, llbb)
124+
}
109125
}
110126
}
111127

0 commit comments

Comments
 (0)