diff --git a/paddle/fluid/framework/data_feed.cu b/paddle/fluid/framework/data_feed.cu index 63e0ce5a5e71e..9bd23c94d96b9 100644 --- a/paddle/fluid/framework/data_feed.cu +++ b/paddle/fluid/framework/data_feed.cu @@ -26,8 +26,6 @@ limitations under the License. */ #include "paddle/fluid/framework/fleet/heter_ps/gpu_graph_node.h" #include "paddle/fluid/framework/fleet/heter_ps/graph_gpu_wrapper.h" -DECLARE_int32(batch_num); - namespace paddle { namespace framework { @@ -349,15 +347,7 @@ int GraphDataGenerator::FillInsBuf() { } -int times = 0; int GraphDataGenerator::GenerateBatch() { - times += 1; - VLOG(0) << "Begin batch " << times; - if (times > FLAGS_batch_num) { - VLOG(0) << "close batch"; - return 0; - } - platform::CUDADeviceGuard guard(gpuid_); int res = 0; while (ins_buf_pair_len_ < batch_size_) { diff --git a/paddle/fluid/platform/flags.cc b/paddle/fluid/platform/flags.cc index f2b5b2907c049..054a804e6b38e 100644 --- a/paddle/fluid/platform/flags.cc +++ b/paddle/fluid/platform/flags.cc @@ -88,25 +88,6 @@ PADDLE_DEFINE_EXPORTED_bool( "input and output must be half precision) and recurrent neural networks " "(RNNs)."); -/** - * CUDA related FLAG - * Name: FLAGS_selected_gpus - * Since Version: 1.3.0 - * Value Range: integer list separated by comma, default empty list - * Example: FLAGS_selected_gpus=0,1,2,3,4,5,6,7 to train or predict with 0~7 gpu - * cards - * Note: A list of device ids separated by comma, like: 0,1,2,3 - */ -PADDLE_DEFINE_EXPORTED_int32( - batch_num, 0, - "A list of device ids separated by comma, like: 0,1,2,3. " - "This option is useful when doing multi process training and " - "each process have only one device (GPU). If you want to use " - "all visible devices, set this to empty string. NOTE: the " - "reason of doing this is that we want to use P2P communication" - "between GPU devices, use CUDA_VISIBLE_DEVICES can only use" - "share-memory only."); - /** * CUDA related FLAG * Name: FLAGS_selected_gpus