File tree Expand file tree Collapse file tree 3 files changed +3
-4
lines changed
Expand file tree Collapse file tree 3 files changed +3
-4
lines changed Original file line number Diff line number Diff line change 55
66#include < cuda_runtime_api.h>
77
8+ // TODO: replace TA_CUDA_CHECK with STD_CUDA_CHECK after
9+ // https://github.com/pytorch/pytorch/pull/169385 has landed.
810#define TA_CUDA_CHECK (...) __VA_ARGS__
911
1012namespace libtorchaudio ::cuda {
Original file line number Diff line number Diff line change @@ -210,12 +210,10 @@ void forced_align_impl(
210210 C10_CUDA_KERNEL_LAUNCH_CHECK ();
211211 ++backPtrBufferLen;
212212 if (backPtrBufferLen == kBackPtrBufferSize || t == T - 1 ) {
213- // cpuDataTranferStream.synchronize();
214213 libtorchaudio::cuda::synchronize (cpuDataTranferStream, device_index);
215214 // GPU -> GPU copy
216215 bufferCopy = torch::stable::clone (backPtrBuffer);
217216 STD_TORCH_CHECK (bufferCopy.is_contiguous (), " unexpected fail, need to implement stable::Tensor::contiguous()" )
218- // defaultStream.synchronize();
219217 libtorchaudio::cuda::synchronize (defaultStream, device_index);
220218 libtorchaudio::cuda::setCurrentCUDAStream (cpuDataTranferStream, device_index);
221219 // Copy ASYNC from GPU to CPU
@@ -231,7 +229,6 @@ void forced_align_impl(
231229 backPtrBufferLen = 0 ;
232230 }
233231 }
234- // cpuDataTranferStream.synchronize();
235232 libtorchaudio::cuda::synchronize (cpuDataTranferStream, device_index);
236233 auto alphasCpu = torchaudio::stable::cpu (alphas);
237234 auto alphasCpu_a = torchaudio::accessor<scalar_t , 2 >(alphasCpu);
Original file line number Diff line number Diff line change 11#pragma once
22// TODO: remove this file once https://github.com/pytorch/pytorch/pull/169376
3- // has landed.
3+ // has landed in nightly .
44
55#include <c10/cuda/CUDAStream.h>
66#include <torch/csrc/inductor/aoti_torch/utils.h>
You can’t perform that action at this time.
0 commit comments