You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I notice that the block size in gpu-latency is 64 rather than 32 (warp size) or a single thread. Is there any consideration to set as 64? Looking forward to your reply:)
The text was updated successfully, but these errors were encountered:
A single thread would work just the same. Using a full warp just feels better, and a full warp is 64 threads on CDNA hardware. On NVIDIA, this actually runs two warps instead of just one, but from what I remember, the interference is low.
Hi, I notice that the block size in gpu-latency is 64 rather than 32 (warp size) or a single thread. Is there any consideration to set as 64? Looking forward to your reply:)
The text was updated successfully, but these errors were encountered: