From 2a184456c88c79271efeb58e9065375cfd610e22 Mon Sep 17 00:00:00 2001 From: Cliff Burdick <30670611+cliffburdick@users.noreply.github.com> Date: Wed, 5 Feb 2025 09:18:01 -0800 Subject: [PATCH] Update fft_cuda.h --- include/matx/transforms/fft/fft_cuda.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/include/matx/transforms/fft/fft_cuda.h b/include/matx/transforms/fft/fft_cuda.h index c2eaa9c2..8f53a041 100644 --- a/include/matx/transforms/fft/fft_cuda.h +++ b/include/matx/transforms/fft/fft_cuda.h @@ -216,7 +216,7 @@ template class matxCUDAFFTPlan_t if (i.IsContiguous() && o.IsContiguous()) { // Previously we used cudaMemGetInfo to get free memory to determine batch size. This can be very slow, // and for small FFTs this call can create extra latency. For now we'll just assume the user knows what - // they're doing and not try to batch FFTs that are too small + // they're doing and not try to batch FFTs that are too large const auto shape = i.Shape(); params.batch = std::accumulate(std::begin(shape), std::end(shape) - 1, 1, std::multiplies()); params.batch_dims = i.Rank() - 1;