What's Changed
- registered hook correctly by @bryce13950 in #1051
- optimized QKV bridge a bit by @bryce13950 in #1046
- Add support for layer norm and bias folding by @degenfabian in #1044
- updated get params to fill zeroes when needed by @bryce13950 in #1049
- Match device selection of TransformerBridge to HookedTransformer by @degenfabian in #1047
- Improve TransformerBridge hook compatibility with HookedTransformers by @degenfabian in #1054
- Enable setting cached hooks by @degenfabian in #1048
- Create bridge for every module in Phi 1 by @degenfabian in #1055
- Rename Neo bridges to be in line with new naming scheme by @degenfabian in #1056
- Rename Mixtral bridges to be in line with new naming scheme by @degenfabian in #1057
- added test and made sure backwards hooks are working by @bryce13950 in #1058
- Remove second layer norm from phi component mapping by @degenfabian in #1059
- Create bridge for every module in pythia by @degenfabian in #1060
- Create bridge for every module in Qwen 2 by @degenfabian in #1061
- Processing functions by @bryce13950 in #1053
- Attempted Processing match by @bryce13950 in #1063
- Process restoration by @bryce13950 in #1064
- Add missing configuration parameters by @degenfabian in #1065
- Properly set up normalization_type and layer_norm_folding attributes in initialized components by @degenfabian in #1066
- Process accuracy by @bryce13950 in #1067
- Ablation hugging face weights by @bryce13950 in #1070
- Ci fixes by @bryce13950 in #1072
- Revision extra forwards by @bryce13950 in #1073
- Test coverage by @bryce13950 in #1074
- Attention hooks full coverage for folding by @bryce13950 in #1078
- Ci job splitting by @bryce13950 in #1079
- fixed batch dimension by @bryce13950 in #1082
- fixed cache hooks by @bryce13950 in #1083
- fixed bias displaying by @bryce13950 in #1084
- fixed return type none by @bryce13950 in #1085
- Create pass through for hooks in compatibility mode by @bryce13950 in #1086
- fixed alias hook props by @bryce13950 in #1087
- made all hooks show properly by @bryce13950 in #1088
- updated loading in main demo to use transformers bridge by @bryce13950 in #1010
- switch from poetry to uv by @mivanit in #1037
- addded full kv cache by @bryce13950 in #1089
- Added full hook coverage for previous keys by @bryce13950 in #988
- updated loading in arena content demo to use transformer bridge by @degenfabian in #1012
- regeneerated with new hooks by @bryce13950 in #1091
- added test coverage for ensuring compatibility by @bryce13950 in #989
- Test hook shape coverage by @bryce13950 in #1000
- Hook compatibility by @bryce13950 in #1092
- Final compatibility coverage by @bryce13950 in #1090
- tested llama 3.1 by @bryce13950 in #1096
- fixed stop at layer by @bryce13950 in #1100
- Duplicate hook fix by @bryce13950 in #1098
- Gemma2 fix by @bryce13950 in #1099
- Fix gpt oss by @bryce13950 in #1101
- created benchmark suite by @bryce13950 in #1104
- finalized t5 adapter by @bryce13950 in #1095
- Model improvements by @bryce13950 in #1105
- decoupling weight processing completely from hooked transformer by @bryce13950 in #1103
- removed invalid comparison by @bryce13950 in #1107
- Revert "decoupling weight processing completely from hooked transformer" by @bryce13950 in #1108
- finalized bench mark logic by @bryce13950 in #1109
- Fix opt by @bryce13950 in #1106
- Benchmarking and compatibility only by @bryce13950 in #1112
- Decouple weight procesing by @bryce13950 in #1114
- optimized benchmarks a bit by @bryce13950 in #1115
- fixed tensor storing by @bryce13950 in #1116
- added skip condition by @bryce13950 in #1117
- Gpt2 weight match by @bryce13950 in #1118
- Gemma3 match by @bryce13950 in #1119
- setup real aliases by @bryce13950 in #1121
- Gpt oss match by @bryce13950 in #1120
- trimmed memory a bit by @bryce13950 in #1122
- created benchmark suite for unsupported models in hooked transformer by @bryce13950 in #1123
- fixed remaining gemma 3 benchmarks by @bryce13950 in #1124
- Gated MLP bridge by @bryce13950 in #1110
- setup brenchmark suite, and trimmed out extra tests by @bryce13950 in #1125
- Attention cleanup by @bryce13950 in #1126
- Benchmarking cross comparison revision by @bryce13950 in #1127
- Oss match by @bryce13950 in #1128
- Cleanup by @bryce13950 in #1129
- Weight processing generalization by @bryce13950 in #1131
- Processing cleanup by @bryce13950 in #1132
- Final cleanup by @bryce13950 in #1135
- Supported Architectures – code artifact cleanup by @jlarson4 in #1136
- Qwen3 adapter by @bryce13950 in #1138
- Model Bridge – Source Keys Cleanup by @jlarson4 in #1137
- cleaned up a lot of things by @bryce13950 in #1113
- Transformer bridge layer norm folding by @bryce13950 in #1071
- Updated release workflow by @bryce13950 in #1146
New Contributors
Full Changelog: v3.0.0a8...v3.0.0b1