-
Notifications
You must be signed in to change notification settings - Fork 172
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
replace triton.ops dependencies in pytorch/ao (#1250)
Summary: Pull Request resolved: #1250 `triton.ops` is moved to kernels directory with the 3.2 update. This change updates imports to be through explicit matmul and matmul_perf_model helper files copied to `pytorch/ao`. Differential Revision: D65678605
- Loading branch information
1 parent
e41ca4e
commit c5b07d7
Showing
7 changed files
with
631 additions
and
7 deletions.
There are no files selected for viewing
Empty file.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,367 @@ | ||
import torch | ||
|
||
from triton import Config, autotune, cdiv, heuristics, jit | ||
from triton import language as tl | ||
from .matmul_perf_model import early_config_prune, estimate_matmul_time | ||
|
||
_ordered_datatypes = [torch.int8, torch.float16, torch.bfloat16, torch.float32] | ||
|
||
|
||
def upcast_if_fp8(a): | ||
if "fp8" in str(a): | ||
return torch.float16 | ||
return a | ||
|
||
|
||
def get_higher_dtype(a, b): | ||
a = upcast_if_fp8(a) | ||
b = upcast_if_fp8(b) | ||
if a is b: | ||
return a | ||
|
||
assert a in _ordered_datatypes | ||
assert b in _ordered_datatypes | ||
|
||
for d in _ordered_datatypes: | ||
if a is d: | ||
return b | ||
if b is d: | ||
return a | ||
|
||
|
||
def init_to_zero(name): | ||
return lambda nargs: nargs[name].zero_() | ||
|
||
|
||
def get_configs_io_bound(): | ||
configs = [] | ||
for num_stages in [2, 3, 4, 5, 6]: | ||
for block_m in [16, 32]: | ||
for block_k in [32, 64]: | ||
for block_n in [32, 64, 128, 256]: | ||
num_warps = 2 if block_n <= 64 else 4 | ||
configs.append( | ||
Config( | ||
{ | ||
"BLOCK_M": block_m, | ||
"BLOCK_N": block_n, | ||
"BLOCK_K": block_k, | ||
"SPLIT_K": 1, | ||
}, | ||
num_stages=num_stages, | ||
num_warps=num_warps, | ||
) | ||
) | ||
# split_k | ||
for split_k in [2, 4, 8, 16]: | ||
configs.append( | ||
Config( | ||
{ | ||
"BLOCK_M": block_m, | ||
"BLOCK_N": block_n, | ||
"BLOCK_K": block_k, | ||
"SPLIT_K": split_k, | ||
}, | ||
num_stages=num_stages, | ||
num_warps=num_warps, | ||
pre_hook=init_to_zero("C"), | ||
) | ||
) | ||
return configs | ||
|
||
|
||
@autotune( | ||
configs=[ | ||
# basic configs for compute-bound matmuls | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 256, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=3, | ||
num_warps=8, | ||
), | ||
Config( | ||
{"BLOCK_M": 256, "BLOCK_N": 128, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=3, | ||
num_warps=8, | ||
), | ||
Config( | ||
{"BLOCK_M": 256, "BLOCK_N": 64, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 256, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 128, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 64, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 128, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 32, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 32, "BLOCK_K": 32, "SPLIT_K": 1}, | ||
num_stages=5, | ||
num_warps=2, | ||
), | ||
# good for int8 | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 256, "BLOCK_K": 128, "SPLIT_K": 1}, | ||
num_stages=3, | ||
num_warps=8, | ||
), | ||
Config( | ||
{"BLOCK_M": 256, "BLOCK_N": 128, "BLOCK_K": 128, "SPLIT_K": 1}, | ||
num_stages=3, | ||
num_warps=8, | ||
), | ||
Config( | ||
{"BLOCK_M": 256, "BLOCK_N": 64, "BLOCK_K": 128, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 256, "BLOCK_K": 128, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 128, "BLOCK_K": 128, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 64, "BLOCK_K": 64, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 128, "BLOCK_K": 64, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 128, "BLOCK_N": 32, "BLOCK_K": 64, "SPLIT_K": 1}, | ||
num_stages=4, | ||
num_warps=4, | ||
), | ||
Config( | ||
{"BLOCK_M": 64, "BLOCK_N": 32, "BLOCK_K": 64, "SPLIT_K": 1}, | ||
num_stages=5, | ||
num_warps=2, | ||
), | ||
] | ||
+ get_configs_io_bound(), | ||
key=["M", "N", "K"], | ||
prune_configs_by={ | ||
"early_config_prune": early_config_prune, | ||
"perf_model": estimate_matmul_time, | ||
"top_k": 10, | ||
}, | ||
) | ||
@heuristics( | ||
{ | ||
"EVEN_K": lambda args: args["K"] % (args["BLOCK_K"] * args["SPLIT_K"]) == 0, | ||
} | ||
) | ||
@jit | ||
def _kernel( | ||
A, | ||
B, | ||
C, | ||
M, | ||
N, | ||
K, # | ||
stride_am, | ||
stride_ak, # | ||
stride_bk, | ||
stride_bn, # | ||
stride_cm, | ||
stride_cn, # | ||
acc_dtype: tl.constexpr, # | ||
input_precision: tl.constexpr, # | ||
fp8_fast_accum: tl.constexpr, # | ||
BLOCK_M: tl.constexpr, | ||
BLOCK_N: tl.constexpr, | ||
BLOCK_K: tl.constexpr, # | ||
GROUP_M: tl.constexpr, | ||
SPLIT_K: tl.constexpr, | ||
EVEN_K: tl.constexpr, | ||
AB_DTYPE: tl.constexpr, # | ||
): | ||
# matrix multiplication | ||
pid = tl.program_id(0) | ||
pid_z = tl.program_id(1) | ||
grid_m = tl.cdiv(M, BLOCK_M) | ||
grid_n = tl.cdiv(N, BLOCK_N) | ||
# re-order program ID for better L2 performance | ||
width = GROUP_M * grid_n | ||
group_id = pid // width | ||
group_size = min(grid_m - group_id * GROUP_M, GROUP_M) | ||
pid_m = group_id * GROUP_M + (pid % group_size) | ||
pid_n = (pid % width) // (group_size) | ||
# do matrix multiplication | ||
rm = pid_m * BLOCK_M + tl.arange(0, BLOCK_M) | ||
rn = pid_n * BLOCK_N + tl.arange(0, BLOCK_N) | ||
ram = tl.max_contiguous(tl.multiple_of(rm % M, BLOCK_M), BLOCK_M) | ||
rbn = tl.max_contiguous(tl.multiple_of(rn % N, BLOCK_N), BLOCK_N) | ||
rk = pid_z * BLOCK_K + tl.arange(0, BLOCK_K) | ||
# pointers | ||
A = A + (ram[:, None] * stride_am + rk[None, :] * stride_ak) | ||
B = B + (rk[:, None] * stride_bk + rbn[None, :] * stride_bn) | ||
acc = tl.zeros((BLOCK_M, BLOCK_N), dtype=acc_dtype) | ||
for k in range(0, tl.cdiv(K, BLOCK_K * SPLIT_K)): | ||
if EVEN_K: | ||
a = tl.load(A) | ||
b = tl.load(B) | ||
else: | ||
k_remaining = K - k * (BLOCK_K * SPLIT_K) | ||
_0 = tl.zeros((1, 1), dtype=C.dtype.element_ty) | ||
a = tl.load(A, mask=rk[None, :] < k_remaining, other=_0) | ||
b = tl.load(B, mask=rk[:, None] < k_remaining, other=_0) | ||
if AB_DTYPE is not None: | ||
a = a.to(AB_DTYPE) | ||
b = b.to(AB_DTYPE) | ||
if fp8_fast_accum: | ||
acc = tl.dot( | ||
a, b, acc, out_dtype=acc_dtype, input_precision=input_precision | ||
) | ||
else: | ||
acc += tl.dot(a, b, out_dtype=acc_dtype, input_precision=input_precision) | ||
A += BLOCK_K * SPLIT_K * stride_ak | ||
B += BLOCK_K * SPLIT_K * stride_bk | ||
acc = acc.to(C.dtype.element_ty) | ||
# rematerialize rm and rn to save registers | ||
rm = pid_m * BLOCK_M + tl.arange(0, BLOCK_M) | ||
rn = pid_n * BLOCK_N + tl.arange(0, BLOCK_N) | ||
C = C + (rm[:, None] * stride_cm + rn[None, :] * stride_cn) | ||
mask = (rm < M)[:, None] & (rn < N)[None, :] | ||
# handles write-back with reduction-splitting | ||
if SPLIT_K == 1: | ||
tl.store(C, acc, mask=mask) | ||
else: | ||
tl.atomic_add(C, acc, mask=mask) | ||
|
||
|
||
class _matmul(torch.autograd.Function): | ||
kernel = _kernel | ||
|
||
_locks = {} | ||
|
||
@staticmethod | ||
def _call(a, b, acc_dtype, input_precision, fp8_fast_accum, output_dtype): | ||
device = a.device | ||
# handle non-contiguous inputs if necessary | ||
if a.stride(0) > 1 and a.stride(1) > 1: | ||
a = a.contiguous() | ||
if b.stride(0) > 1 and b.stride(1) > 1: | ||
b = b.contiguous() | ||
# checks constraints | ||
assert ( | ||
a.shape[1] == b.shape[0] | ||
), f"incompatible dimensions {a.shape} and {b.shape}" | ||
M, K = a.shape | ||
_, N = b.shape | ||
|
||
# common type between a and b | ||
ab_dtype = get_higher_dtype(a.dtype, b.dtype) | ||
|
||
# allocates output | ||
if output_dtype is None: | ||
output_dtype = ab_dtype | ||
|
||
c = torch.empty((M, N), device=device, dtype=output_dtype) | ||
|
||
# Allowed types for acc_type given the types of a and b. | ||
supported_acc_dtypes = { | ||
torch.float16: (torch.float32, torch.float16), | ||
torch.bfloat16: (torch.float32, torch.bfloat16), | ||
torch.float32: (torch.float32,), | ||
torch.int8: (torch.int32,), | ||
} | ||
|
||
if acc_dtype is None: | ||
acc_dtype = supported_acc_dtypes[ab_dtype][0] | ||
else: | ||
assert isinstance(acc_dtype, torch.dtype), "acc_dtype must be a torch.dtype" | ||
assert ( | ||
acc_dtype in supported_acc_dtypes[a.dtype] | ||
), "acc_dtype not compatible with the type of a" | ||
assert ( | ||
acc_dtype in supported_acc_dtypes[b.dtype] | ||
), "acc_dtype not compatible with the type of b" | ||
|
||
def to_tl_type(ty): | ||
return getattr(tl, str(ty).split(".")[-1]) | ||
|
||
acc_dtype = to_tl_type(acc_dtype) | ||
ab_dtype = to_tl_type(ab_dtype) | ||
output_dtype = to_tl_type(output_dtype) | ||
|
||
# Tensor cores support input with mixed float8 types. | ||
if a.dtype in [tl.float8e4nv, tl.float8e5] and b.dtype in [ | ||
tl.float8e4nv, | ||
tl.float8e5, | ||
]: | ||
ab_dtype = None | ||
# launch kernel | ||
grid = lambda META: ( | ||
cdiv(M, META["BLOCK_M"]) * cdiv(N, META["BLOCK_N"]), | ||
META["SPLIT_K"], | ||
) | ||
_kernel[grid]( | ||
a, | ||
b, | ||
c, | ||
M, | ||
N, | ||
K, # | ||
a.stride(0), | ||
a.stride(1), # | ||
b.stride(0), | ||
b.stride(1), # | ||
c.stride(0), | ||
c.stride(1), # | ||
acc_dtype=acc_dtype, # | ||
input_precision=input_precision, # | ||
fp8_fast_accum=fp8_fast_accum, # | ||
GROUP_M=8, | ||
AB_DTYPE=ab_dtype, | ||
) | ||
return c | ||
|
||
@staticmethod | ||
def forward( | ||
ctx, | ||
a, | ||
b, | ||
acc_dtype=None, | ||
input_precision=None, | ||
fp8_fast_accum=True, | ||
output_dtype=None, | ||
): | ||
return _matmul._call( | ||
a, | ||
b, | ||
acc_dtype=acc_dtype, | ||
input_precision=input_precision, | ||
fp8_fast_accum=fp8_fast_accum, | ||
output_dtype=output_dtype, | ||
) | ||
|
||
|
||
matmul = _matmul.apply |
Oops, something went wrong.