Skip to content

Commit f5ed0cb

Browse files
committed
[RISCV] Add target feature to force-enable atomics
This adds a +forced-atomics target feature with the same semantics as +atomics-32 on ARM (D130480). For RISCV targets without the +a extension, this forces LLVM to assume that lock-free atomics (up to 32/64 bits for riscv32/64 respectively) are available. This means that atomic load/store are lowered to a simple load/store (and fence as necessary), as these are guaranteed to be atomic (as long as they're aligned). Atomic RMW/CAS are lowered to __sync (rather than __atomic) libcalls. Responsibility for providing the __sync libcalls lies with the user (for privileged single-core code they can be implemented by disabling interrupts). Code using +forced-atomics and -forced-atomics are not ABI compatible if atomic variables cross the ABI boundary. For context, the difference between __sync and __atomic is that the former are required to be lock-free, while the latter requires a shared global lock provided by a shared object library. See https://llvm.org/docs/Atomics.html#libcalls-atomic for a detailed discussion on the topic. This target feature will be used by Rust's riscv32i target family to support the use of atomic load/store without atomic RMW/CAS. Differential Revision: https://reviews.llvm.org/D130621
1 parent 6da3f90 commit f5ed0cb

File tree

6 files changed

+3697
-17
lines changed

6 files changed

+3697
-17
lines changed

llvm/docs/Atomics.rst

+7
Original file line numberDiff line numberDiff line change
@@ -581,6 +581,13 @@ case. The only common architecture without that property is SPARC -- SPARCV8 SMP
581581
systems were common, yet it doesn't support any sort of compare-and-swap
582582
operation.
583583

584+
Some targets (like RISCV) support a ``+forced-atomics`` target feature, which
585+
enables the use of lock-free atomics even if LLVM is not aware of any specific
586+
OS support for them. In this case, the user is responsible for ensuring that
587+
necessary ``__sync_*`` implementations are available. Code using
588+
``+forced-atomics`` is ABI-incompatible with code not using the feature, if
589+
atomic variables cross the ABI boundary.
590+
584591
In either of these cases, the Target in LLVM can claim support for atomics of an
585592
appropriate size, and then implement some subset of the operations via libcalls
586593
to a ``__sync_*`` function. Such functions *must* not use locks in their

llvm/lib/Target/RISCV/RISCV.td

+11
Original file line numberDiff line numberDiff line change
@@ -481,6 +481,17 @@ def TuneSiFive7 : SubtargetFeature<"sifive7", "RISCVProcFamily", "SiFive7",
481481
"SiFive 7-Series processors",
482482
[TuneNoDefaultUnroll]>;
483483

484+
// Assume that lock-free native-width atomics are available, even if the target
485+
// and operating system combination would not usually provide them. The user
486+
// is responsible for providing any necessary __sync implementations. Code
487+
// built with this feature is not ABI-compatible with code built without this
488+
// feature, if atomic variables are exposed across the ABI boundary.
489+
def FeatureForcedAtomics : SubtargetFeature<
490+
"forced-atomics", "HasForcedAtomics", "true",
491+
"Assume that lock-free native-width atomics are available">;
492+
def HasAtomicLdSt
493+
: Predicate<"Subtarget->hasStdExtA() || Subtarget->hasForcedAtomics()">;
494+
484495
//===----------------------------------------------------------------------===//
485496
// Named operands for CSR instructions.
486497
//===----------------------------------------------------------------------===//

llvm/lib/Target/RISCV/RISCVISelLowering.cpp

+20
Original file line numberDiff line numberDiff line change
@@ -411,6 +411,8 @@ RISCVTargetLowering::RISCVTargetLowering(const TargetMachine &TM,
411411
if (Subtarget.hasStdExtA()) {
412412
setMaxAtomicSizeInBitsSupported(Subtarget.getXLen());
413413
setMinCmpXchgSizeInBits(32);
414+
} else if (Subtarget.hasForcedAtomics()) {
415+
setMaxAtomicSizeInBitsSupported(Subtarget.getXLen());
414416
} else {
415417
setMaxAtomicSizeInBitsSupported(0);
416418
}
@@ -929,6 +931,16 @@ RISCVTargetLowering::RISCVTargetLowering(const TargetMachine &TM,
929931
}
930932
}
931933

934+
if (Subtarget.hasForcedAtomics()) {
935+
// Set atomic rmw/cas operations to expand to force __sync libcalls.
936+
setOperationAction(
937+
{ISD::ATOMIC_CMP_SWAP, ISD::ATOMIC_SWAP, ISD::ATOMIC_LOAD_ADD,
938+
ISD::ATOMIC_LOAD_SUB, ISD::ATOMIC_LOAD_AND, ISD::ATOMIC_LOAD_OR,
939+
ISD::ATOMIC_LOAD_XOR, ISD::ATOMIC_LOAD_NAND, ISD::ATOMIC_LOAD_MIN,
940+
ISD::ATOMIC_LOAD_MAX, ISD::ATOMIC_LOAD_UMIN, ISD::ATOMIC_LOAD_UMAX},
941+
XLenVT, Expand);
942+
}
943+
932944
// Function alignments.
933945
const Align FunctionAlignment(Subtarget.hasStdExtC() ? 2 : 4);
934946
setMinFunctionAlignment(FunctionAlignment);
@@ -12286,6 +12298,10 @@ RISCVTargetLowering::shouldExpandAtomicRMWInIR(AtomicRMWInst *AI) const {
1228612298
if (AI->isFloatingPointOperation())
1228712299
return AtomicExpansionKind::CmpXChg;
1228812300

12301+
// Don't expand forced atomics, we want to have __sync libcalls instead.
12302+
if (Subtarget.hasForcedAtomics())
12303+
return AtomicExpansionKind::None;
12304+
1228912305
unsigned Size = AI->getType()->getPrimitiveSizeInBits();
1229012306
if (Size == 8 || Size == 16)
1229112307
return AtomicExpansionKind::MaskedIntrinsic;
@@ -12389,6 +12405,10 @@ Value *RISCVTargetLowering::emitMaskedAtomicRMWIntrinsic(
1238912405
TargetLowering::AtomicExpansionKind
1239012406
RISCVTargetLowering::shouldExpandAtomicCmpXchgInIR(
1239112407
AtomicCmpXchgInst *CI) const {
12408+
// Don't expand forced atomics, we want to have __sync libcalls instead.
12409+
if (Subtarget.hasForcedAtomics())
12410+
return AtomicExpansionKind::None;
12411+
1239212412
unsigned Size = CI->getCompareOperand()->getType()->getPrimitiveSizeInBits();
1239312413
if (Size == 8 || Size == 16)
1239412414
return AtomicExpansionKind::MaskedIntrinsic;

llvm/lib/Target/RISCV/RISCVInstrInfoA.td

+15-17
Original file line numberDiff line numberDiff line change
@@ -105,20 +105,25 @@ defm AMOMAXU_D : AMO_rr_aq_rl<0b11100, 0b011, "amomaxu.d">,
105105
// Pseudo-instructions and codegen patterns
106106
//===----------------------------------------------------------------------===//
107107

108-
let Predicates = [HasStdExtA] in {
109-
110-
/// Atomic loads and stores
111-
108+
// Atomic load/store are available under both +a and +force-atomics.
112109
// Fences will be inserted for atomic load/stores according to the logic in
113110
// RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}.
111+
let Predicates = [HasAtomicLdSt] in {
112+
defm : LdPat<atomic_load_8, LB>;
113+
defm : LdPat<atomic_load_16, LH>;
114+
defm : LdPat<atomic_load_32, LW>;
115+
116+
defm : AtomicStPat<atomic_store_8, SB, GPR>;
117+
defm : AtomicStPat<atomic_store_16, SH, GPR>;
118+
defm : AtomicStPat<atomic_store_32, SW, GPR>;
119+
}
114120

115-
defm : LdPat<atomic_load_8, LB>;
116-
defm : LdPat<atomic_load_16, LH>;
117-
defm : LdPat<atomic_load_32, LW>;
121+
let Predicates = [HasAtomicLdSt, IsRV64] in {
122+
defm : LdPat<atomic_load_64, LD, i64>;
123+
defm : AtomicStPat<atomic_store_64, SD, GPR, i64>;
124+
}
118125

119-
defm : AtomicStPat<atomic_store_8, SB, GPR>;
120-
defm : AtomicStPat<atomic_store_16, SH, GPR>;
121-
defm : AtomicStPat<atomic_store_32, SW, GPR>;
126+
let Predicates = [HasStdExtA] in {
122127

123128
/// AMOs
124129

@@ -304,13 +309,6 @@ def : Pat<(int_riscv_masked_cmpxchg_i32
304309

305310
let Predicates = [HasStdExtA, IsRV64] in {
306311

307-
/// 64-bit atomic loads and stores
308-
309-
// Fences will be inserted for atomic load/stores according to the logic in
310-
// RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}.
311-
defm : LdPat<atomic_load_64, LD, i64>;
312-
defm : AtomicStPat<atomic_store_64, SD, GPR, i64>;
313-
314312
defm : AMOPat<"atomic_swap_64", "AMOSWAP_D">;
315313
defm : AMOPat<"atomic_load_add_64", "AMOADD_D">;
316314
defm : AMOPat<"atomic_load_and_64", "AMOAND_D">;

llvm/lib/Target/RISCV/RISCVSubtarget.h

+2
Original file line numberDiff line numberDiff line change
@@ -98,6 +98,7 @@ class RISCVSubtarget : public RISCVGenSubtargetInfo {
9898
bool EnableSaveRestore = false;
9999
bool EnableUnalignedScalarMem = false;
100100
bool HasLUIADDIFusion = false;
101+
bool HasForcedAtomics = false;
101102
unsigned XLen = 32;
102103
unsigned ZvlLen = 0;
103104
MVT XLenVT = MVT::i32;
@@ -194,6 +195,7 @@ class RISCVSubtarget : public RISCVGenSubtargetInfo {
194195
bool enableSaveRestore() const { return EnableSaveRestore; }
195196
bool enableUnalignedScalarMem() const { return EnableUnalignedScalarMem; }
196197
bool hasLUIADDIFusion() const { return HasLUIADDIFusion; }
198+
bool hasForcedAtomics() const { return HasForcedAtomics; }
197199
MVT getXLenVT() const { return XLenVT; }
198200
unsigned getXLen() const { return XLen; }
199201
unsigned getFLen() const {

0 commit comments

Comments
 (0)