-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathbenchmarking_offline.py
More file actions
161 lines (131 loc) · 5.68 KB
/
benchmarking_offline.py
File metadata and controls
161 lines (131 loc) · 5.68 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
import argparse
import gc
import json
import math
import os
import random
import time
import pandas
import torch
from benchmarking_utils import run_all_perf_funcs_once
from tools.utils import build_from_sources, extract_bm_bk_bn
torch.set_grad_enabled(False)
print("=====================Benchmarking Script -- Offline Mode======================")
parser = argparse.ArgumentParser()
parser.add_argument("--mnk", type=str, required=True)
parser.add_argument("--warmup_seconds", type=float, required=True)
parser.add_argument("--benchmark_seconds", type=float, required=True)
parser.add_argument("--base_dir", type=str, required=True)
parser.add_argument("--gpu_device_id", type=int, required=True)
parser.add_argument("--perf_func", type=str, required=True)
args = parser.parse_args()
torch.set_grad_enabled(False)
load_start = time.time()
hgemm = build_from_sources(mnk=args.mnk, base_dir=args.base_dir, verbose=False)
load_end = time.time()
print(f"Load hgemm module time: {load_end - load_start:.2f} seconds")
def main():
torch.cuda.set_device(args.gpu_device_id)
mnk = args.mnk
m, n, k = map(int, mnk.split("_"))
warmup_seconds = args.warmup_seconds
benchmark_seconds = args.benchmark_seconds
torch.cuda.synchronize()
start_time = time.time()
print(f"m={m}, n={n}, k={k}, Warmup={warmup_seconds}s, Benchmark={benchmark_seconds}s")
torch.cuda.synchronize()
hgemm.init_cublas_handle() # type: ignore
hgemm.init_cublaslt_handle_v1() # type: ignore
hgemm.init_cublaslt_handle_v2() # type: ignore
torch.cuda.synchronize()
if args.perf_func in "hgemm_cublaslt_auto_tuning_tn":
print("Finding best algo for hgemm_cublaslt_auto_tuning_tn...")
start_time = time.time()
hgemm.find_best_algo_tn_v2_torch(m, n, k) # type: ignore
torch.cuda.synchronize()
end_time = time.time()
print(f"Find best algo time: {end_time - start_time:.2f} seconds")
elif args.perf_func in "hgemm_cublaslt_auto_tuning_nn":
print("Finding best algo for hgemm_cublaslt_auto_tuning_nn...")
start_time = time.time()
hgemm.find_best_algo_nn_v2_torch(m, n, k) # type: ignore
torch.cuda.synchronize()
end_time = time.time()
print(f"Find best algo time: {end_time - start_time:.2f} seconds")
perf_func_name_to_func = {
"hgemm_cublas_tn": hgemm.hgemm_cublas_tn, # type: ignore
"hgemm_cublas_nn": hgemm.hgemm_cublas_nn, # type: ignore
"hgemm_cublaslt_heuristic_tn": hgemm.hgemm_cublaslt_heuristic_tn, # type: ignore
"hgemm_cublaslt_heuristic_nn": hgemm.hgemm_cublaslt_heuristic_nn, # type: ignore
"hgemm_cublaslt_auto_tuning_tn": hgemm.hgemm_cublaslt_auto_tuning_tn, # type: ignore
"hgemm_cublaslt_auto_tuning_nn": hgemm.hgemm_cublaslt_auto_tuning_nn, # type: ignore
"matmul": torch.matmul,
}
perf_func_list = [
perf_func_name_to_func[args.perf_func],
hgemm.cuda_l2_a100_fp16, # type: ignore
]
origin_perf_func_list = perf_func_list.copy()
with open(f"kernels/a100_F16F16F16F16/{mnk}.cu", "r") as f:
code_text = f.read()
bm, bk, bn = extract_bm_bk_bn(code_text)
if bm > 0 and bk > 0 and bn > 0:
padding_m = math.ceil(m / bm) * bm - m
padding_k = math.ceil(k / bk) * bk - k
padding_n = math.ceil(n / bn) * bn - n
else:
padding_m, padding_k, padding_n = 0, 0, 0
print(f"Using padding_m={padding_m}, padding_k={padding_k}, padding_n={padding_n}")
# Warmup
print("Warmup...")
warmup_start_time = time.time()
warmup_count = 0
while time.time()- warmup_start_time < warmup_seconds:
record = run_all_perf_funcs_once(
perf_func_list=perf_func_list, m=m, n=n, k=k,
padding_m=padding_m, padding_k=padding_k, padding_n=padding_n
)
warmup_count += 1
print(f"Warmup done: {warmup_count} iterations in {time.time() - warmup_start_time:.2f} seconds.")
print("Benchmarking...")
records = []
benchmark_start_time = time.time()
benchmark_count = 0
while time.time() - benchmark_start_time < benchmark_seconds:
random.shuffle(perf_func_list)
record = run_all_perf_funcs_once(
perf_func_list=perf_func_list, m=m, n=n, k=k,
padding_m=padding_m, padding_k=padding_k, padding_n=padding_n
)
record["idx"] = benchmark_count
records.append(record)
benchmark_count += 1
hgemm.destroy_cublas_handle() # type: ignore
hgemm.destroy_cublaslt_handle_v1() # type: ignore
hgemm.destroy_cublaslt_handle_v2() # type: ignore
torch.cuda.synchronize()
gc.collect()
torch.cuda.empty_cache()
gc.collect()
end_time = time.time()
print(f"Total time: {(end_time - start_time):.2f} seconds, {len(records)} records collected.")
# Print results
func_names = [func.__name__ for func in origin_perf_func_list]
func_names_ms = [func.__name__ + "_ms" for func in origin_perf_func_list]
df = pandas.DataFrame.from_records(records, columns=["idx"] + func_names + func_names_ms)
print(df.head().to_markdown())
print(df.tail().to_markdown())
mean_tflops = df[func_names].mean()
merged_result = mean_tflops.to_dict()
mean_ms = df[func_names_ms].mean()
merged_result["version"] = "202511261845"
print(merged_result)
print(mean_ms)
our_speed = mean_tflops["cuda_l2_a100_fp16"]
baseline_speed = mean_tflops[args.perf_func]
print(f"speedup over {args.perf_func}: {our_speed / baseline_speed:.2f}x")
with open(os.path.join(args.base_dir, f"benchmark_result_{args.perf_func}.json"), "w") as f:
json.dump({"records": merged_result}, f)
if __name__ == "__main__":
main()
torch.cuda.empty_cache()