Skip to content

Commit

Permalink
[ci test] ignore backward
Browse files Browse the repository at this point in the history
  • Loading branch information
gouzil committed Apr 25, 2024
1 parent 4bcbcc6 commit 5378583
Showing 1 changed file with 56 additions and 56 deletions.
112 changes: 56 additions & 56 deletions test/cpp/prim/test_eager_prim.cc
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@
#include "paddle/common/flags.h"
#include "paddle/fluid/eager/api/generated/eager_generated/forwards/dygraph_functions.h"
#include "paddle/fluid/eager/api/utils/hook_utils.h"
#include "paddle/fluid/eager/backward.h"
// #include "paddle/fluid/eager/backward.h"
#include "paddle/fluid/prim/utils/utils.h"
#include "paddle/phi/core/dense_tensor.h"
#include "paddle/phi/core/kernel_registry.h"
Expand All @@ -32,62 +32,62 @@ COMMON_DECLARE_string(tensor_operants_mode);
namespace paddle {
namespace prim {

TEST(EagerPrim, TanhBackwardTest) {
// 1. Initialized
eager_test::InitEnv(paddle::platform::CPUPlace());
FLAGS_tensor_operants_mode = "eager";
paddle::prim::InitTensorOperants();
// 2. pre
paddle::framework::DDim ddim = common::make_ddim({4, 16, 16, 32});
paddle::Tensor tensor0 =
eager_test::CreateTensorWithValue(ddim,
paddle::platform::CPUPlace(),
phi::DataType::FLOAT32,
phi::DataLayout::NCHW,
5.0 /*value*/,
true /*is_leaf*/);
::egr::egr_utils_api::RetainGradForTensor(tensor0);
paddle::Tensor tensor1 =
eager_test::CreateTensorWithValue(ddim,
paddle::platform::CPUPlace(),
phi::DataType::FLOAT32,
phi::DataLayout::NCHW,
5.0 /*value*/,
true /*is_leaf*/);
::egr::egr_utils_api::RetainGradForTensor(tensor1);
// 3. Run Forward once
paddle::Tensor out0 = tanh_ad_func(tensor0);
std::vector<paddle::Tensor> outs0 = {out0};
// Disable prim
PrimCommonUtils::SetBwdPrimEnabled(false);
ASSERT_FALSE(PrimCommonUtils::IsBwdPrimEnabled());
// 4. Run Backward
egr::Backward(outs0, {}, false);
// TEST(EagerPrim, TanhBackwardTest) {
// // 1. Initialized
// eager_test::InitEnv(paddle::platform::CPUPlace());
// FLAGS_tensor_operants_mode = "eager";
// paddle::prim::InitTensorOperants();
// // 2. pre
// paddle::framework::DDim ddim = common::make_ddim({4, 16, 16, 32});
// paddle::Tensor tensor0 =
// eager_test::CreateTensorWithValue(ddim,
// paddle::platform::CPUPlace(),
// phi::DataType::FLOAT32,
// phi::DataLayout::NCHW,
// 5.0 /*value*/,
// true /*is_leaf*/);
// ::egr::egr_utils_api::RetainGradForTensor(tensor0);
// paddle::Tensor tensor1 =
// eager_test::CreateTensorWithValue(ddim,
// paddle::platform::CPUPlace(),
// phi::DataType::FLOAT32,
// phi::DataLayout::NCHW,
// 5.0 /*value*/,
// true /*is_leaf*/);
// ::egr::egr_utils_api::RetainGradForTensor(tensor1);
// // 3. Run Forward once
// paddle::Tensor out0 = tanh_ad_func(tensor0);
// std::vector<paddle::Tensor> outs0 = {out0};
// // Disable prim
// PrimCommonUtils::SetBwdPrimEnabled(false);
// ASSERT_FALSE(PrimCommonUtils::IsBwdPrimEnabled());
// // 4. Run Backward
// egr::Backward(outs0, {}, false);

paddle::Tensor out1 = tanh_ad_func(tensor1);
std::vector<paddle::Tensor> outs1 = {out1};
// Enable prim
PrimCommonUtils::SetBwdPrimEnabled(true);
ASSERT_TRUE(PrimCommonUtils::IsBwdPrimEnabled());
// 4. Run Backward
::egr::Backward(outs1, {}, false);
VLOG(7)
<< "Target Grad is: "
<< std::static_pointer_cast<phi::DenseTensor>(
::egr::EagerUtils::unsafe_autograd_meta(tensor0)->Grad().impl())
->data<float>()[0];
VLOG(7)
<< "Result Grad is: "
<< std::static_pointer_cast<phi::DenseTensor>(
::egr::EagerUtils::unsafe_autograd_meta(tensor1)->Grad().impl())
->data<float>()[0];
// Examine Backward Grad
eager_test::CompareGradTensorWithValue<float>(
tensor1,
std::static_pointer_cast<phi::DenseTensor>(
::egr::EagerUtils::unsafe_autograd_meta(tensor0)->Grad().impl())
->data<float>()[0]);
}
// paddle::Tensor out1 = tanh_ad_func(tensor1);
// std::vector<paddle::Tensor> outs1 = {out1};
// // Enable prim
// PrimCommonUtils::SetBwdPrimEnabled(true);
// ASSERT_TRUE(PrimCommonUtils::IsBwdPrimEnabled());
// // 4. Run Backward
// ::egr::Backward(outs1, {}, false);
// VLOG(7)
// << "Target Grad is: "
// << std::static_pointer_cast<phi::DenseTensor>(
// ::egr::EagerUtils::unsafe_autograd_meta(tensor0)->Grad().impl())
// ->data<float>()[0];
// VLOG(7)
// << "Result Grad is: "
// << std::static_pointer_cast<phi::DenseTensor>(
// ::egr::EagerUtils::unsafe_autograd_meta(tensor1)->Grad().impl())
// ->data<float>()[0];
// // Examine Backward Grad
// eager_test::CompareGradTensorWithValue<float>(
// tensor1,
// std::static_pointer_cast<phi::DenseTensor>(
// ::egr::EagerUtils::unsafe_autograd_meta(tensor0)->Grad().impl())
// ->data<float>()[0]);
// }

TEST(EagerPrim, LogicalOperantsTest) {
// 1. Initialized
Expand Down

0 comments on commit 5378583

Please sign in to comment.