@@ -832,53 +832,6 @@ std::tuple<Tensor,Tensor> _th_gels(const Tensor & self, const Tensor & A) {
832832 }
833833 return std::tuple<Tensor, Tensor>(res1, res2);
834834}
835- std::tuple<Tensor &,Tensor &> _th_eig_out (Tensor & res1, Tensor & res2, const Tensor & self, bool eigenvectors) {
836- // DeviceGuard omitted
837- auto dispatch_scalar_type = infer_scalar_type (self);
838-
839- switch (dispatch_scalar_type) {
840- case ScalarType::Double: {
841- auto res1_ = checked_dense_tensor_unwrap (res1, " res1" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
842- auto res2_ = checked_dense_tensor_unwrap (res2, " res2" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
843- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
844- THDoubleTensor_geev (res1_, res2_, self_, eigenvectors);
845- break ;
846- }
847- case ScalarType::Float: {
848- auto res1_ = checked_dense_tensor_unwrap (res1, " res1" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
849- auto res2_ = checked_dense_tensor_unwrap (res2, " res2" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
850- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
851- THFloatTensor_geev (res1_, res2_, self_, eigenvectors);
852- break ;
853- }
854- default :
855- AT_ERROR (" _th_eig_out not supported on CPUType for " , dispatch_scalar_type);
856- }
857- return std::tuple<Tensor &, Tensor &>(res1, res2);
858- }
859- std::tuple<Tensor,Tensor> _th_eig (const Tensor & self, bool eigenvectors) {
860- // DeviceGuard omitted
861- auto dispatch_scalar_type = infer_scalar_type (self);
862- auto res1_ = c10::make_intrusive<TensorImpl, UndefinedTensorImpl>(c10::Storage (c10::Storage::use_byte_size_t (), 0 , allocator (), true ),DispatchKey::CPU, scalarTypeToTypeMeta (dispatch_scalar_type)).release ();
863- auto res1 = Tensor (c10::intrusive_ptr<TensorImpl, UndefinedTensorImpl>::reclaim (res1_));
864- auto res2_ = c10::make_intrusive<TensorImpl, UndefinedTensorImpl>(c10::Storage (c10::Storage::use_byte_size_t (), 0 , allocator (), true ),DispatchKey::CPU, scalarTypeToTypeMeta (dispatch_scalar_type)).release ();
865- auto res2 = Tensor (c10::intrusive_ptr<TensorImpl, UndefinedTensorImpl>::reclaim (res2_));
866- switch (dispatch_scalar_type) {
867- case ScalarType::Double: {
868- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig" , false , DeviceType::CPU, dispatch_scalar_type);
869- THDoubleTensor_geev (res1_, res2_, self_, eigenvectors);
870- break ;
871- }
872- case ScalarType::Float: {
873- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig" , false , DeviceType::CPU, dispatch_scalar_type);
874- THFloatTensor_geev (res1_, res2_, self_, eigenvectors);
875- break ;
876- }
877- default :
878- AT_ERROR (" _th_eig not supported on CPUType for " , dispatch_scalar_type);
879- }
880- return std::tuple<Tensor, Tensor>(res1, res2);
881- }
882835Tensor & _th_potri_out (Tensor & output, const Tensor & self, bool upper) {
883836 // DeviceGuard omitted
884837 auto dispatch_scalar_type = infer_scalar_type (self);
0 commit comments