You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -270,6 +271,45 @@ Overhead of `log_forward_pass` vs raw `model.forward()`. See `test_outputs/repor
270
271
271
272
*Overhead is dominated by per-operation bookkeeping. Large models with fewer, heavier ops (VGG16) show lower relative overhead. Small models with many lightweight ops show higher relative overhead. All measurements on CPU.*
272
273
274
+
### Decoration Overhead (logging disabled)
275
+
276
+
TorchLens permanently wraps all ~2000 PyTorch functions at import time. When logging
277
+
is disabled, each wrapper is a single bool check (`if not _logging_enabled: return func(...)`).
*Heavy ops (conv2d, matmul, SDPA) show <1% overhead — within measurement noise. The ~600ns wrapper cost is only visible on sub-10μs elementwise ops. In practice, decoration has negligible impact on model inference speed.*
292
+
293
+
---
294
+
295
+
## Large Graph Scaling
296
+
297
+
TorchLens supports visualization of very large computational graphs using the ELK layout engine
298
+
(auto-selected above 3,500 nodes, or via `vis_node_placement="elk"`).
0 commit comments