|
210 | 210 | variants: function |
211 | 211 | dispatch: |
212 | 212 | CUDA: fused_dropout_cuda |
| 213 | + tags: nondeterministic_seeded |
213 | 214 |
|
214 | 215 | - func: _masked_scale(Tensor self, Tensor mask, float scale) -> Tensor |
215 | 216 | variants: function |
|
221 | 222 | dispatch: |
222 | 223 | CPU: native_dropout_cpu |
223 | 224 | CUDA: native_dropout_cuda |
| 225 | + tags: nondeterministic_seeded |
224 | 226 |
|
225 | 227 | - func: native_dropout_backward(Tensor grad_output, Tensor mask, float scale) -> Tensor |
226 | 228 | dispatch: |
|
243 | 245 | dispatch: |
244 | 246 | CompositeImplicitAutograd: dropout |
245 | 247 | NestedTensorCPU, NestedTensorCUDA: dropout_nested |
| 248 | + tags: nondeterministic_seeded |
246 | 249 |
|
247 | 250 | - func: dropout_(Tensor(a!) self, float p, bool train) -> Tensor(a!) |
248 | 251 | dispatch: |
|
892 | 895 | variants: function, method |
893 | 896 | dispatch: |
894 | 897 | CompositeExplicitAutograd: bernoulli |
| 898 | + tags: nondeterministic_seeded |
895 | 899 |
|
896 | 900 | - func: bernoulli.out(Tensor self, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!) |
897 | 901 | device_check: NoCheck # TensorIterator |
|
924 | 928 | - func: bernoulli.p(Tensor self, float p, *, Generator? generator=None) -> Tensor |
925 | 929 | device_check: NoCheck # TensorIterator |
926 | 930 | variants: function, method |
| 931 | + tags: nondeterministic_seeded |
927 | 932 |
|
928 | 933 | - func: bilinear(Tensor input1, Tensor input2, Tensor weight, Tensor? bias=None) -> Tensor |
929 | 934 |
|
|
3741 | 3746 | device_guard: False |
3742 | 3747 |
|
3743 | 3748 | - func: rand(int[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
| 3749 | + tags: nondeterministic_seeded |
3744 | 3750 |
|
3745 | 3751 | - func: rand.generator(int[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
3746 | 3752 |
|
|
3749 | 3755 | - func: rand.generator_out(int[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!) |
3750 | 3756 |
|
3751 | 3757 | - func: rand_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor |
| 3758 | + tags: nondeterministic_seeded |
3752 | 3759 |
|
3753 | 3760 | - func: randint(int high, int[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
| 3761 | + tags: nondeterministic_seeded |
3754 | 3762 |
|
3755 | 3763 | - func: randint.generator(int high, int[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
3756 | 3764 |
|
3757 | 3765 | - func: randint.low(int low, int high, int[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
| 3766 | + tags: nondeterministic_seeded |
3758 | 3767 |
|
3759 | 3768 | - func: randint.low_generator(int low, int high, int[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
3760 | 3769 |
|
|
3767 | 3776 | - func: randint.low_generator_out(int low, int high, int[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!) |
3768 | 3777 |
|
3769 | 3778 | - func: randint_like(Tensor self, int high, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor |
| 3779 | + tags: nondeterministic_seeded |
3770 | 3780 |
|
3771 | 3781 | - func: randint_like.low_dtype(Tensor self, int low, int high, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor |
| 3782 | + tags: nondeterministic_seeded |
3772 | 3783 |
|
3773 | 3784 | - func: randn(int[] size, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
| 3785 | + tags: nondeterministic_seeded |
3774 | 3786 |
|
3775 | 3787 | - func: randn.generator(int[] size, *, Generator? generator, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
3776 | 3788 |
|
|
3787 | 3799 | - func: randn.generator_out(int[] size, *, Generator? generator, Tensor(a!) out) -> Tensor(a!) |
3788 | 3800 |
|
3789 | 3801 | - func: randn_like(Tensor self, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None, MemoryFormat? memory_format=None) -> Tensor |
| 3802 | + tags: nondeterministic_seeded |
3790 | 3803 |
|
3791 | 3804 | - func: randperm(int n, *, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
| 3805 | + tags: nondeterministic_seeded |
3792 | 3806 |
|
3793 | 3807 | - func: randperm.generator(int n, *, Generator? generator, ScalarType? dtype=long, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
3794 | 3808 |
|
|
3957 | 3971 |
|
3958 | 3972 | - func: rrelu(Tensor self, Scalar lower=0.125, Scalar upper=0.3333333333333333, bool training=False, Generator? generator=None) -> Tensor |
3959 | 3973 | device_check: NoCheck # TensorIterator |
| 3974 | + tags: nondeterministic_seeded |
3960 | 3975 |
|
3961 | 3976 | - func: rrelu_(Tensor(a!) self, Scalar lower=0.125, Scalar upper=0.3333333333333333, bool training=False, Generator? generator=None) -> Tensor(a!) |
3962 | 3977 | device_check: NoCheck # TensorIterator |
|
5186 | 5201 | dispatch: |
5187 | 5202 | CPU: _s_gamma_cpu |
5188 | 5203 | CUDA: _s_gamma_cuda |
| 5204 | + tags: nondeterministic_seeded |
5189 | 5205 |
|
5190 | 5206 | - func: _dirichlet_grad(Tensor x, Tensor alpha, Tensor total) -> Tensor |
5191 | 5207 | dispatch: |
|
5203 | 5219 | dispatch: |
5204 | 5220 | CPU: _s_poisson_cpu |
5205 | 5221 | CUDA: _s_poisson_cuda |
| 5222 | + tags: nondeterministic_seeded |
5206 | 5223 |
|
5207 | 5224 | - func: binomial(Tensor count, Tensor prob, Generator? generator=None) -> Tensor |
5208 | 5225 | device_check: NoCheck # TensorIterator |
5209 | 5226 | dispatch: |
5210 | 5227 | CPU: _s_binomial_cpu |
5211 | 5228 | CUDA: _s_binomial_cuda |
| 5229 | + tags: nondeterministic_seeded |
5212 | 5230 |
|
5213 | 5231 | # When more variants get ported to native, this dispatch will get more |
5214 | 5232 | # complicated |
|
7756 | 7774 | variants: method, function |
7757 | 7775 | dispatch: |
7758 | 7776 | CPU, CUDA: multinomial |
| 7777 | + tags: nondeterministic_seeded |
7759 | 7778 |
|
7760 | 7779 | - func: lgamma.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!) |
7761 | 7780 | device_check: NoCheck # TensorIterator |
|
8446 | 8465 | CPU, CUDA: normal |
8447 | 8466 | MPS: normal_mps |
8448 | 8467 | Meta: normal_meta |
| 8468 | + tags: nondeterministic_seeded |
8449 | 8469 |
|
8450 | 8470 | - func: normal.float_Tensor_out(float mean, Tensor std, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!) |
8451 | 8471 | dispatch: |
|
8458 | 8478 | CPU, CUDA: normal |
8459 | 8479 | MPS: normal_mps |
8460 | 8480 | Meta: normal_meta |
| 8481 | + tags: nondeterministic_seeded |
8461 | 8482 |
|
8462 | 8483 | - func: normal.Tensor_Tensor_out(Tensor mean, Tensor std, *, Generator? generator=None, Tensor(a!) out) -> Tensor(a!) |
8463 | 8484 | dispatch: |
|
8470 | 8491 | CPU, CUDA: normal |
8471 | 8492 | MPS: normal_mps |
8472 | 8493 | Meta: normal_meta |
| 8494 | + tags: nondeterministic_seeded |
8473 | 8495 |
|
8474 | 8496 | - func: normal.float_float(float mean, float std, int[] size, *, Generator? generator=None, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor |
8475 | 8497 |
|
|
9710 | 9732 | dispatch: |
9711 | 9733 | CPU: rrelu_with_noise_cpu |
9712 | 9734 | CUDA: rrelu_with_noise_cuda |
| 9735 | + tags: nondeterministic_seeded |
9713 | 9736 |
|
9714 | 9737 | - func: rrelu_with_noise_backward(Tensor grad_output, Tensor self, Tensor noise, Scalar lower, Scalar upper, bool training, bool self_is_result) -> Tensor |
9715 | 9738 | python_module: nn |
|
0 commit comments