@@ -1615,118 +1615,6 @@ internal Avx10v1() { }
1615
1615
/// </summary>
1616
1616
public static Vector256 < ulong > Min ( Vector256 < ulong > left , Vector256 < ulong > right ) => Min ( left , right ) ;
1617
1617
1618
- /// <summary>
1619
- /// __m128 _mm_fmadd_ps (__m128 a, __m128 b, __m128 c)
1620
- /// VFMADDPS xmm1, xmm2, xmm3/m128
1621
- /// VFMADDPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1622
- /// </summary>
1623
- public static Vector128 < float > MultiplyAdd ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplyAdd ( a , b , c ) ;
1624
-
1625
- /// <summary>
1626
- /// __m128d _mm_fmadd_pd (__m128d a, __m128d b, __m128d c)
1627
- /// VFMADDPD xmm1, xmm2, xmm3/m128
1628
- /// VFMADDPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1629
- /// </summary>
1630
- public static Vector128 < double > MultiplyAdd ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplyAdd ( a , b , c ) ;
1631
-
1632
- /// <summary>
1633
- /// __m256 _mm256_fmadd_ps (__m256 a, __m256 b, __m256 c)
1634
- /// VFMADDPS ymm1, ymm2, ymm3/m256
1635
- /// VFMADDPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1636
- /// </summary>
1637
- public static Vector256 < float > MultiplyAdd ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplyAdd ( a , b , c ) ;
1638
-
1639
- /// <summary>
1640
- /// __m256d _mm256_fmadd_pd (__m256d a, __m256d b, __m256d c)
1641
- /// VFMADDPD ymm1, ymm2, ymm3/m256
1642
- /// VFMADDPD ymm1 {k1}{z}, ymm2, ymm3/m256/m64bcst
1643
- /// </summary>
1644
- public static Vector256 < double > MultiplyAdd ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplyAdd ( a , b , c ) ;
1645
-
1646
- /// <summary>
1647
- /// __m128 _mm_fnmadd_ps (__m128 a, __m128 b, __m128 c)
1648
- /// VFNMADDPS xmm1, xmm2, xmm3/m128
1649
- /// VFNMADDPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1650
- /// </summary>
1651
- public static Vector128 < float > MultiplyAddNegated ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplyAddNegated ( a , b , c ) ;
1652
-
1653
- /// <summary>
1654
- /// __m128d _mm_fnmadd_pd (__m128d a, __m128d b, __m128d c)
1655
- /// VFNMADDPD xmm1, xmm2, xmm3/m128
1656
- /// VFNMADDPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1657
- /// </summary>
1658
- public static Vector128 < double > MultiplyAddNegated ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplyAddNegated ( a , b , c ) ;
1659
-
1660
- /// <summary>
1661
- /// __m256 _mm256_fnmadd_ps (__m256 a, __m256 b, __m256 c)
1662
- /// VFNMADDPS ymm1, ymm2, ymm3/m256
1663
- /// VFNMADDPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1664
- /// </summary>
1665
- public static Vector256 < float > MultiplyAddNegated ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplyAddNegated ( a , b , c ) ;
1666
-
1667
- /// <summary>
1668
- /// __m256d _mm256_fnmadd_pd (__m256d a, __m256d b, __m256d c)
1669
- /// VFNMADDPD ymm1, ymm2, ymm3/m256
1670
- /// VFNMADDPD ymm1 {k1}{z}, ymm2, ymm3/m256/m64bcst
1671
- /// </summary>
1672
- public static Vector256 < double > MultiplyAddNegated ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplyAddNegated ( a , b , c ) ;
1673
-
1674
- /// <summary>
1675
- /// __m128 _mm_fnmadd_ss (__m128 a, __m128 b, __m128 c)
1676
- /// VFNMADDSS xmm1, xmm2, xmm3/m32
1677
- /// VFNMADDSS xmm1 {k1}{z}, xmm2, xmm3/m32{er}
1678
- /// </summary>
1679
- public static Vector128 < float > MultiplyAddNegatedScalar ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplyAddNegatedScalar ( a , b , c ) ;
1680
-
1681
- /// <summary>
1682
- /// __m128d _mm_fnmadd_sd (__m128d a, __m128d b, __m128d c)
1683
- /// VFNMADDSD xmm1, xmm2, xmm3/m64
1684
- /// VFNMADDSD xmm1 {k1}{z}, xmm2, xmm3/m64{er}
1685
- /// </summary>
1686
- public static Vector128 < double > MultiplyAddNegatedScalar ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplyAddNegatedScalar ( a , b , c ) ;
1687
-
1688
- /// <summary>
1689
- /// __m128 _mm_fmadd_ss (__m128 a, __m128 b, __m128 c)
1690
- /// VFMADDSS xmm1, xmm2, xmm3/m32
1691
- /// VFMADDSS xmm1 {k1}{z}, xmm2, xmm3/m32{er}
1692
- /// </summary>
1693
- public static Vector128 < float > MultiplyAddScalar ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplyAddScalar ( a , b , c ) ;
1694
-
1695
- /// <summary>
1696
- /// __m128d _mm_fmadd_sd (__m128d a, __m128d b, __m128d c)
1697
- /// VFMADDSD xmm1, xmm2, xmm3/m64
1698
- /// VFMADDSD xmm1 {k1}{z}, xmm2, xmm3/m64{er}
1699
- /// </summary>
1700
- public static Vector128 < double > MultiplyAddScalar ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplyAddScalar ( a , b , c ) ;
1701
-
1702
- /// <summary>
1703
- /// __m128 _mm_fmaddsub_ps (__m128 a, __m128 b, __m128 c)
1704
- /// VFMADDSUBPS xmm1, xmm2, xmm3/m128
1705
- /// VFMADDSUBPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1706
- /// </summary>
1707
- public static Vector128 < float > MultiplyAddSubtract ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplyAddSubtract ( a , b , c ) ;
1708
-
1709
- /// <summary>
1710
- /// __m128d _mm_fmaddsub_pd (__m128d a, __m128d b, __m128d c)
1711
- /// VFMADDSUBPD xmm1, xmm2, xmm3/m128
1712
- /// VFMADDSUBPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1713
- /// </summary>
1714
- public static Vector128 < double > MultiplyAddSubtract ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplyAddSubtract ( a , b , c ) ;
1715
-
1716
- /// <summary>
1717
- /// __m256 _mm256_fmaddsub_ps (__m256 a, __m256 b, __m256 c)
1718
- /// VFMADDSUBPS ymm1, ymm2, ymm3/m256
1719
- /// VFMADDSUBPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1720
- /// </summary>
1721
- public static Vector256 < float > MultiplyAddSubtract ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplyAddSubtract ( a , b , c ) ;
1722
-
1723
- /// <summary>
1724
- /// __m256d _mm256_fmaddsub_pd (__m256d a, __m256d b, __m256d c)
1725
- /// VFMADDSUBPD ymm1, ymm2, ymm3/m256
1726
- /// VFMADDSUBPD ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1727
- /// </summary>
1728
- public static Vector256 < double > MultiplyAddSubtract ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplyAddSubtract ( a , b , c ) ;
1729
-
1730
1618
/// <summary>
1731
1619
/// __m128i _mm_mullo_epi64 (__m128i a, __m128i b)
1732
1620
/// VPMULLQ xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
@@ -1763,118 +1651,6 @@ internal Avx10v1() { }
1763
1651
/// </summary>
1764
1652
public static Vector128 < float > MultiplyScalar ( Vector128 < float > left , Vector128 < float > right , [ ConstantExpected ( Max = FloatRoundingMode . ToZero ) ] FloatRoundingMode mode ) => MultiplyScalar ( left , right , mode ) ;
1765
1653
1766
- /// <summary>
1767
- /// __m128 _mm_fmsub_ps (__m128 a, __m128 b, __m128 c)
1768
- /// VFMSUBPS xmm1, xmm2, xmm3/m128
1769
- /// VFMSUBPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1770
- /// </summary>
1771
- public static Vector128 < float > MultiplySubtract ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplySubtract ( a , b , c ) ;
1772
-
1773
- /// <summary>
1774
- /// __m128d _mm_fmsub_pd (__m128d a, __m128d b, __m128d c)
1775
- /// VFMSUBPD xmm1, xmm2, xmm3/m128
1776
- /// VFMSUBPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1777
- /// </summary>
1778
- public static Vector128 < double > MultiplySubtract ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplySubtract ( a , b , c ) ;
1779
-
1780
- /// <summary>
1781
- /// __m256 _mm256_fmsub_ps (__m256 a, __m256 b, __m256 c)
1782
- /// VFMSUBPS ymm1, ymm2, ymm3/m256
1783
- /// VFMSUBPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1784
- /// </summary>
1785
- public static Vector256 < float > MultiplySubtract ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplySubtract ( a , b , c ) ;
1786
-
1787
- /// <summary>
1788
- /// __m256d _mm256_fmsub_pd (__m256d a, __m256d b, __m256d c)
1789
- /// VFMSUBPD ymm1, ymm2, ymm3/m256
1790
- /// VFMSUBPD ymm1 {k1}{z}, ymm2, ymm3/m256/m64bcst
1791
- /// </summary>
1792
- public static Vector256 < double > MultiplySubtract ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplySubtract ( a , b , c ) ;
1793
-
1794
- /// <summary>
1795
- /// __m128 _mm_fmsubadd_ps (__m128 a, __m128 b, __m128 c)
1796
- /// VFMSUBADDPS xmm1, xmm2, xmm3/m128
1797
- /// VFMSUBADDPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1798
- /// </summary>
1799
- public static Vector128 < float > MultiplySubtractAdd ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplySubtractAdd ( a , b , c ) ;
1800
-
1801
- /// <summary>
1802
- /// __m128d _mm_fmsubadd_pd (__m128d a, __m128d b, __m128d c)
1803
- /// VFMSUBADDPD xmm1, xmm2, xmm3/m128
1804
- /// VFMSUBADDPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1805
- /// </summary>
1806
- public static Vector128 < double > MultiplySubtractAdd ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplySubtractAdd ( a , b , c ) ;
1807
-
1808
- /// <summary>
1809
- /// __m256 _mm256_fmsubadd_ps (__m256 a, __m256 b, __m256 c)
1810
- /// VFMSUBADDPS ymm1, ymm2, ymm3/m256
1811
- /// VFMSUBADDPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1812
- /// </summary>
1813
- public static Vector256 < float > MultiplySubtractAdd ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplySubtractAdd ( a , b , c ) ;
1814
-
1815
- /// <summary>
1816
- /// __m256d _mm256_fmsubadd_pd (__m256d a, __m256d b, __m256d c)
1817
- /// VFMSUBADDPD ymm1, ymm2, ymm3/m256
1818
- /// VFMSUBADDPD ymm1 {k1}{z}, ymm2, ymm3/m256/m64bcst
1819
- /// </summary>
1820
- public static Vector256 < double > MultiplySubtractAdd ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplySubtractAdd ( a , b , c ) ;
1821
-
1822
- /// <summary>
1823
- /// __m128 _mm_fnmsub_ps (__m128 a, __m128 b, __m128 c)
1824
- /// VFNMSUBPS xmm1, xmm2, xmm3/m128
1825
- /// VFNMSUBPS xmm1 {k1}{z}, xmm2, xmm3/m128/m32bcst
1826
- /// </summary>
1827
- public static Vector128 < float > MultiplySubtractNegated ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplySubtractNegated ( a , b , c ) ;
1828
-
1829
- /// <summary>
1830
- /// __m128d _mm_fnmsub_pd (__m128d a, __m128d b, __m128d c)
1831
- /// VFNMSUBPD xmm1, xmm2, xmm3/m128
1832
- /// VFNMSUBPD xmm1 {k1}{z}, xmm2, xmm3/m128/m64bcst
1833
- /// </summary>
1834
- public static Vector128 < double > MultiplySubtractNegated ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplySubtractNegated ( a , b , c ) ;
1835
-
1836
- /// <summary>
1837
- /// __m256 _mm256_fnmsub_ps (__m256 a, __m256 b, __m256 c)
1838
- /// VFNMSUBPS ymm1, ymm2, ymm3/m256
1839
- /// VFNMSUBPS ymm1 {k1}{z}, ymm2, ymm3/m256/m32bcst
1840
- /// </summary>
1841
- public static Vector256 < float > MultiplySubtractNegated ( Vector256 < float > a , Vector256 < float > b , Vector256 < float > c ) => MultiplySubtractNegated ( a , b , c ) ;
1842
-
1843
- /// <summary>
1844
- /// __m256d _mm256_fnmsub_pd (__m256d a, __m256d b, __m256d c)
1845
- /// VFNMSUBPD ymm1, ymm2, ymm3/m256
1846
- /// VFNMSUBPD ymm1 {k1}{z}, ymm2, ymm3/m256/m64bcst
1847
- /// </summary>
1848
- public static Vector256 < double > MultiplySubtractNegated ( Vector256 < double > a , Vector256 < double > b , Vector256 < double > c ) => MultiplySubtractNegated ( a , b , c ) ;
1849
-
1850
- /// <summary>
1851
- /// __m128 _mm_fnmsub_ss (__m128 a, __m128 b, __m128 c)
1852
- /// VFNMSUBSS xmm1, xmm2, xmm3/m32
1853
- /// VFNMSUBSS xmm1 {k1}{z}, xmm2, xmm3/m32{er}
1854
- /// </summary>
1855
- public static Vector128 < float > MultiplySubtractNegatedScalar ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplySubtractNegatedScalar ( a , b , c ) ;
1856
-
1857
- /// <summary>
1858
- /// __m128d _mm_fnmsub_sd (__m128d a, __m128d b, __m128d c)
1859
- /// VFNMSUBSD xmm1, xmm2, xmm3/m64
1860
- /// VFNMSUBSD xmm1 {k1}{z}, xmm2, xmm3/m64{er}
1861
- /// </summary>
1862
- public static Vector128 < double > MultiplySubtractNegatedScalar ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplySubtractNegatedScalar ( a , b , c ) ;
1863
-
1864
- /// <summary>
1865
- /// __m128d _mm_fmsub_sd (__m128d a, __m128d b, __m128d c)
1866
- /// VFMSUBSD xmm1, xmm2, xmm3/m64
1867
- /// VFMSUBSD xmm1 {k1}{z}, xmm2, xmm3/m64{er}
1868
- /// </summary>
1869
- public static Vector128 < double > MultiplySubtractScalar ( Vector128 < double > a , Vector128 < double > b , Vector128 < double > c ) => MultiplySubtractScalar ( a , b , c ) ;
1870
-
1871
- /// <summary>
1872
- /// __m128 _mm_fmsub_ss (__m128 a, __m128 b, __m128 c)
1873
- /// VFMSUBSS xmm1, xmm2, xmm3/m32
1874
- /// VFMSUBSS xmm1 {k1}{z}, xmm2, xmm3/m32{er}
1875
- /// </summary>
1876
- public static Vector128 < float > MultiplySubtractScalar ( Vector128 < float > a , Vector128 < float > b , Vector128 < float > c ) => MultiplySubtractScalar ( a , b , c ) ;
1877
-
1878
1654
/// <summary>
1879
1655
/// __m256i _mm256_permutevar16x16_epi16 (__m256i a, __m256i b)
1880
1656
/// VPERMW ymm1 {k1}{z}, ymm2, ymm3/m256
0 commit comments