You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* init new test
* test pad vocab size to
* add logs
* log to warning
* change TP
* fix loop
* revert
* remove hack size
* this new test should pass
* test not divisible by num tp
* Revert "remove hack size"
This reverts commit bcc6d8d.
* Revert "Revert "remove hack size""
This reverts commit 8322f89.
* Revert "test not divisible by num tp"
This reverts commit 92614bf.
* Revert "this new test should pass"
This reverts commit 9e17a4f.
* change info to warning
* change to print
* add print
* test 2
* new print
* woups
* more
* woups
* comment
* raise errors
* woups
* pad to save vocab size
* simplify test
* assert test raised
* print error msg
* check msg error
* check error
* woups
* clean
* simplify
* remove unused print
* add comment
* add test multiple of tp size
* add print
* add check
* clean
* Update megatron/mpu/layers.py
Co-authored-by: Thomas Wang <[email protected]>
* Update megatron/tokenizer/tokenizer.py
Co-authored-by: Thomas Wang <[email protected]>
* chnage micro-batch-size
* use tiny vocab
* fix data dir
* fix arg
* change micro-batch-size
* adept input ids
* assertIn
* change micro batch size
* Fix test TP
Co-authored-by: Thomas Wang <[email protected]>
* unused var
* add test make_vocab_size_divisible_by
* fix test_tokenizer_vocab_size_multiple_of_tp_size test
* Fix padded vocab size on preprocessing scripts (#257)
* Add tokenizer options in preprocessing scripts
* This should fix the TP issue?
Co-authored-by: SaulLu <[email protected]>
* documentation
Co-authored-by: Thomas Wang <[email protected]>
raiseValueError(f"There is an input id in the input that is greater than the highest possible input id.\nInput: {input_}\nnum_embeddings: {self.num_embeddings}")
0 commit comments