@@ -8721,7 +8721,7 @@ pub unsafe fn vmull_laneq_u32<const LANE: i32>(a: uint32x2_t, b: uint32x4_t) ->
8721
8721
/// Floating-point fused Multiply-Add to accumulator(vector)
8722
8722
#[inline]
8723
8723
#[target_feature(enable = "neon")]
8724
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8724
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8725
8725
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfma))]
8726
8726
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmla))]
8727
8727
pub unsafe fn vfma_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t) -> float32x2_t {
@@ -8737,7 +8737,7 @@ vfma_f32_(b, c, a)
8737
8737
/// Floating-point fused Multiply-Add to accumulator(vector)
8738
8738
#[inline]
8739
8739
#[target_feature(enable = "neon")]
8740
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8740
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8741
8741
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfma))]
8742
8742
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmla))]
8743
8743
pub unsafe fn vfmaq_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t) -> float32x4_t {
@@ -8753,27 +8753,27 @@ vfmaq_f32_(b, c, a)
8753
8753
/// Floating-point fused Multiply-Add to accumulator(vector)
8754
8754
#[inline]
8755
8755
#[target_feature(enable = "neon")]
8756
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8756
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8757
8757
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfma))]
8758
8758
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmla))]
8759
8759
pub unsafe fn vfma_n_f32(a: float32x2_t, b: float32x2_t, c: f32) -> float32x2_t {
8760
- vfma_f32(a, b, vdup_n_f32 (c))
8760
+ vfma_f32(a, b, vdup_n_f32_vfp4 (c))
8761
8761
}
8762
8762
8763
8763
/// Floating-point fused Multiply-Add to accumulator(vector)
8764
8764
#[inline]
8765
8765
#[target_feature(enable = "neon")]
8766
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8766
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8767
8767
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfma))]
8768
8768
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmla))]
8769
8769
pub unsafe fn vfmaq_n_f32(a: float32x4_t, b: float32x4_t, c: f32) -> float32x4_t {
8770
- vfmaq_f32(a, b, vdupq_n_f32 (c))
8770
+ vfmaq_f32(a, b, vdupq_n_f32_vfp4 (c))
8771
8771
}
8772
8772
8773
8773
/// Floating-point fused multiply-subtract from accumulator
8774
8774
#[inline]
8775
8775
#[target_feature(enable = "neon")]
8776
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8776
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8777
8777
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfms))]
8778
8778
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmls))]
8779
8779
pub unsafe fn vfms_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t) -> float32x2_t {
@@ -8784,7 +8784,7 @@ pub unsafe fn vfms_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t) -> float3
8784
8784
/// Floating-point fused multiply-subtract from accumulator
8785
8785
#[inline]
8786
8786
#[target_feature(enable = "neon")]
8787
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8787
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8788
8788
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfms))]
8789
8789
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmls))]
8790
8790
pub unsafe fn vfmsq_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t) -> float32x4_t {
@@ -8795,21 +8795,21 @@ pub unsafe fn vfmsq_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t) -> float
8795
8795
/// Floating-point fused Multiply-subtract to accumulator(vector)
8796
8796
#[inline]
8797
8797
#[target_feature(enable = "neon")]
8798
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8798
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8799
8799
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfms))]
8800
8800
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmls))]
8801
8801
pub unsafe fn vfms_n_f32(a: float32x2_t, b: float32x2_t, c: f32) -> float32x2_t {
8802
- vfms_f32(a, b, vdup_n_f32 (c))
8802
+ vfms_f32(a, b, vdup_n_f32_vfp4 (c))
8803
8803
}
8804
8804
8805
8805
/// Floating-point fused Multiply-subtract to accumulator(vector)
8806
8806
#[inline]
8807
8807
#[target_feature(enable = "neon")]
8808
- #[cfg_attr(target_arch = "arm", target_feature(enable = "fp-armv8,v8 "))]
8808
+ #[cfg_attr(target_arch = "arm", target_feature(enable = "vfp4 "))]
8809
8809
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vfms))]
8810
8810
#[cfg_attr(all(test, target_arch = "aarch64"), assert_instr(fmls))]
8811
8811
pub unsafe fn vfmsq_n_f32(a: float32x4_t, b: float32x4_t, c: f32) -> float32x4_t {
8812
- vfmsq_f32(a, b, vdupq_n_f32 (c))
8812
+ vfmsq_f32(a, b, vdupq_n_f32_vfp4 (c))
8813
8813
}
8814
8814
8815
8815
/// Subtract
0 commit comments