Home
last modified time | relevance | path

Searched defs:Sm70 (Results 1 – 25 of 36) sorted by relevance

12

/external/pytorch/aten/src/ATen/native/transformers/cuda/mem_eff_attention/kernels/
DcutlassB_f16_aligned_k128.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 128, 64, 128, tr… in __launch_bounds__()
50AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 128, 64, 128>::k… in __launch_bounds__()
107AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 128, tru… in __launch_bounds__()
164AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 128>::kN… in __launch_bounds__()
DcutlassB_f16_notaligned_k65536_dropout.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 128, 64, 65536>:… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 64, 64, 65536>::… in __launch_bounds__()
DcutlassB_f16_notaligned_k128_dropout.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 128, 64, 128>::k… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 64, 64, 128>::kN… in __launch_bounds__()
DcutlassB_f16_notaligned_k128.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 128, 64, 128>::… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 64, 64, 128>::k… in __launch_bounds__()
DcutlassB_f16_notaligned_k65536.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 128, 64, 65536>… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 64, 64, 65536>:… in __launch_bounds__()
DcutlassF_f32_notaligned.cu31 AttentionKernel<float, cutlass::arch::Sm70, false, 64, 64, 64, true, true>::kNumThreads, in __launch_bounds__()
88 AttentionKernel<float, cutlass::arch::Sm70, false, 32, 128, 128, true, true>::kNumThreads, in __launch_bounds__()
145 AttentionKernel<float, cutlass::arch::Sm70, false, 32, 128, 65536, true, true>::kNumThreads, in __launch_bounds__()
DcutlassF_f16_notaligned.cu31 … AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, false, 64, 64, 64, true, true>::kNumThreads, in __launch_bounds__()
88 …AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, false, 32, 128, 128, true, true>::kNumThread… in __launch_bounds__()
145 …AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, false, 32, 128, 65536, true, true>::kNumThre… in __launch_bounds__()
DcutlassB_f16_aligned_k32.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 32, true… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 32>::kNu… in __launch_bounds__()
DcutlassB_f16_aligned_k64.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 64, true… in __launch_bounds__()
69AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 64>::kNu… in __launch_bounds__()
DcutlassB_f16_aligned_k65536_dropout.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, true, false, 128, 64, 65536>::… in __launch_bounds__()
88AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, true, false, 64, 64, 65536>::k… in __launch_bounds__()
DcutlassB_f16_aligned_k128_dropout.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, true, false, 128, 64, 128>::kN… in __launch_bounds__()
88AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, true, false, 64, 64, 128>::kNu… in __launch_bounds__()
DcutlassB_f16_aligned_k65536.cu12AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 128, 64, 65536>:… in __launch_bounds__()
88AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, true, false, false, 64, 64, 65536>::… in __launch_bounds__()
DcutlassF_f16_aligned.cu31 … AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, true, 64, 64, 64, true, true>::kNumThreads, in __launch_bounds__()
107 …AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, true, 32, 128, 128, true, true>::kNumThreads, in __launch_bounds__()
183 …AttentionKernel<cutlass::half_t, cutlass::arch::Sm70, true, 32, 128, 65536, true, true>::kNumThrea… in __launch_bounds__()
DcutlassF_f32_aligned.cu31 AttentionKernel<float, cutlass::arch::Sm70, true, 64, 64, 64, true, true>::kNumThreads, in __launch_bounds__()
107 AttentionKernel<float, cutlass::arch::Sm70, true, 32, 128, 128, true, true>::kNumThreads, in __launch_bounds__()
183 AttentionKernel<float, cutlass::arch::Sm70, true, 32, 128, 65536, true, true>::kNumThreads, in __launch_bounds__()
DcutlassB_f32_notaligned_k65536_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, true, false, 64, 64, 65536>::kNumThread… in __launch_bounds__()
DcutlassB_f16_notaligned_k32.cu31AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 64, 64, 32>::kN… in __launch_bounds__()
DcutlassB_f16_notaligned_k64.cu31AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, false, false, 64, 64, 64>::kN… in __launch_bounds__()
DcutlassB_f32_notaligned_k64_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, true, false, 64, 64, 64>::kNumThreads, in __launch_bounds__()
DcutlassB_f32_notaligned_k128_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, true, false, 64, 64, 128>::kNumThreads, in __launch_bounds__()
DcutlassB_f32_notaligned_k32_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, true, false, 64, 64, 32>::kNumThreads, in __launch_bounds__()
DcutlassB_f32_notaligned_k128.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, false, false, 64, 64, 128>::kNumThreads, in __launch_bounds__()
DcutlassB_f32_notaligned_k65536.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, false, false, 64, 64, 65536>::kNumThrea… in __launch_bounds__()
DcutlassB_f16_notaligned_k32_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 64, 64, 32>::kNu… in __launch_bounds__()
DcutlassB_f32_notaligned_k64.cu31AttentionBackwardKernel<cutlass::arch::Sm70, float, false, false, false, 64, 64, 64>::kNumThreads, in __launch_bounds__()
DcutlassB_f16_notaligned_k64_dropout.cu31AttentionBackwardKernel<cutlass::arch::Sm70, cutlass::half_t, false, true, false, 64, 64, 64>::kNu… in __launch_bounds__()

12