Skip to content

Releases: foundation-model-stack/aiu-fms-testing-utils

v0.5.0

09 Dec 18:29
fc3a30f

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.4.4...v0.5.0

v0.4.4

05 Nov 13:59
3ca15db

Choose a tag to compare

What's Changed

  • Chunked prefill support for paged attention by @ani300 in #122
  • fixed bug with loading validation info outputs by @JRosenkranz in #161

Full Changelog: v0.4.3...v0.4.4

v0.4.3

28 Oct 22:07
150b68f

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.4.2...v0.4.3

v0.4.2

08 Oct 02:56
99e6bd1

Choose a tag to compare

What's Changed

Full Changelog: v0.4.1...v0.4.2

v0.4.1

07 Oct 18:20
abe35d3

Choose a tag to compare

What's Changed

  • [utils] decrement_version fix by @kcirred in #146
  • DPP: check valid prompt shapes until one is available in the dataset by @lupalby in #145
  • added dpp to test_scripts by @JRosenkranz in #147
  • [dpp] store enforce_sizes in log name and added generic kwargs to get_default_validation_prefix by @kcirred in #136
  • update llama model expectations tests by @JRosenkranz in #150

Full Changelog: v0.4.0...v0.4.1

v0.4.0

27 Sep 00:27
ee218a4

Choose a tag to compare

What's Changed

Full Changelog: v0.3.0...v0.4.0

v0.3.0

25 Sep 15:02
4b7f51b

Choose a tag to compare

What's Changed

  • Test decoder long ctx by @kcirred in #117
  • Get criteria from DT artifact by @lupalby in #118
  • open gpt.json read-only to support parallel reading. by @gpaulsen in #126
  • Drive Paged Program Script enhancements by @JRosenkranz in #128
  • [dpp] eliminated pad_token_id from print by @kcirred in #130
  • Add the ability to enforce homogeneous program ids in prefill in DPP script by @JRosenkranz in #131
  • update test scripts to work with 4 layer micro model by @JRosenkranz in #134
  • fixed inference.py for batch size 1 symbolic sdpa by @JRosenkranz in #135
  • Make limits more flexible by @ani300 in #138
  • Allow specific user prompts in DPP script by @JRosenkranz in #139
  • Fix warmup to match vllm by @JRosenkranz in #141
  • Add ability in DPP script to select one or many programs that satisfy min batch and min sequence requirements by @JRosenkranz in #137
  • Fix paged generate with too much padding by @ani300 in #142
  • clean_up_tokenization_spaces=True (default) causes incorrect number of tokens after sampling by @JRosenkranz in #143
  • fixed issue where program_id was int when should have been string by @JRosenkranz in #144

New Contributors

Full Changelog: v0.2.3...v0.3.0

v0.2.3

15 Sep 21:01
046f9c4

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.2.2...v0.2.3

v0.2.2

09 Sep 17:55
178bc89

Choose a tag to compare

What's Changed

  • modification to enforce_size behavior to start accepting samples even before enforcing sizes when there is sufficient space by @kcirred in #121

Full Changelog: v0.2.1...v0.2.2

v0.2.1

05 Sep 19:17
4a74f0a

Choose a tag to compare

What's Changed

  • Add a script which drives particular programs during decode when using paged model by @JRosenkranz in #109
  • Add model testing results to pytest xml report by @pavi2707 in #114
  • Include version in validation info outputs by @JRosenkranz in #119
  • [test_decoder] add aftu version to validation log name and search for closest version log name by @kcirred in #115
  • add truncation option to enforce_sizes to truncate prompts from a larger length to meet sizes that may not be available by @kcirred in #113
  • Update transformers by @joerunde in #120

New Contributors

Full Changelog: v0.2.0...v0.2.1